[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (111)

Search Parameters:
Keywords = omnidirectional image

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 21957 KiB  
Article
Terrain Analysis According to Multiscale Surface Roughness in the Taklimakan Desert
by Sebastiano Trevisani and Peter L. Guth
Land 2024, 13(11), 1843; https://doi.org/10.3390/land13111843 - 5 Nov 2024
Viewed by 702
Abstract
Surface roughness, interpreted in the wide sense of surface texture, is a generic term referring to a variety of aspects and scales of spatial variability of surfaces. The analysis of solid earth surface roughness is useful for understanding, characterizing, and monitoring geomorphic factors [...] Read more.
Surface roughness, interpreted in the wide sense of surface texture, is a generic term referring to a variety of aspects and scales of spatial variability of surfaces. The analysis of solid earth surface roughness is useful for understanding, characterizing, and monitoring geomorphic factors at multiple spatiotemporal scales. The different geomorphic features characterizing a landscape exhibit specific characteristics and scales of surface texture. The capability to selectively analyze specific roughness metrics at multiple spatial scales represents a key tool in geomorphometric analysis. This research presents a simplified geostatistical approach for the multiscale analysis of surface roughness, or of image texture in the case of images, that is highly informative and interpretable. The implemented approach is able to describe two main aspects of short-range surface roughness: omnidirectional roughness and roughness anisotropy. Adopting simple upscaling approaches, it is possible to perform a multiscale analysis of roughness. An overview of the information extraction potential of the approach is shown for the analysis of a portion of the Taklimakan desert (China) using a 30 m resolution DEM derived from the Copernicus Glo-30 DSM. The multiscale roughness indexes are used as input features for unsupervised and supervised learning tasks. The approach can be refined both from the perspective of the multiscale analysis as well as in relation to the surface roughness indexes considered. However, even in its present, simplified form, it can find direct applications in relation to multiple contexts and research topics. Full article
(This article belongs to the Section Land, Soil and Water)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Reprojected COP DEM (30 m resolution, UTM F44) of the area of interest overlaid on the hillshade.</p>
Full article ">Figure 2
<p>Sentinel-2 true color RGB image (bands 4, 3, and 2) of the study area, with the main dune morphologies labeled.</p>
Full article ">Figure 3
<p>Main dune morphologies in the study area, visualized using Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>). From top to bottom: network/transverse dunes, longitudinal and transverse dunes, and dome-shaped dunes.</p>
Full article ">Figure 4
<p>Mixed morphologies in the area of interest, visualized using Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>). From top to bottom: outcropping bedrock with shadow and linear dunes, fluvial morphology, and a flat area with minor dune morphologies.</p>
Full article ">Figure 5
<p>RA direction, where the RA strength is higher than 0.3, overlaid on the hillshade (<b>a</b>) and the residual DEM (<b>b</b>) calculated for level L2.</p>
Full article ">Figure 6
<p>Omnidirectional short-range roughness (m) for the different resolutions. Different color scales for each diagram.</p>
Full article ">Figure 7
<p>Roughness anisotropy strength at different resolutions. Different color scales for each diagram.</p>
Full article ">Figure 8
<p>RGB image (each band normalized) of 3 omnidirectional roughness indexes computed at different resolutions (B = L1; G = L2; R = L4). Despite the high correlation of the three indexes, they differentiate very well the morphological features of the area. For example, they markedly highlight the characteristic smoothness of interdune areas of the longitudinal dunes south of the mountain ridge.</p>
Full article ">Figure 9
<p>RGB image (each band normalized) of 3 anisotropy strength roughness indexes computed at different resolutions (B = L1; G = L4; R = L16). In the dune fields north of the mountains, long-wavelength anisotropic features prevail; in contrast, for the southern longitudinal dunes, shorter anisotropic features (L4) are highlighted.</p>
Full article ">Figure 10
<p>Landscape clustered according to multiscale surface roughness indexes. The cluster centers in terms of OR and RA are described in <a href="#land-13-01843-f011" class="html-fig">Figure 11</a>.</p>
Full article ">Figure 11
<p>Cluster centers of the 7 classes resulting from K-means clustering for OR and RA at the different levels.</p>
Full article ">Figure 12
<p>MRI clustering results in the area of the northern dune field, characterized by network and transverse dunes. Clustering results (<b>d</b>), Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>).</p>
Full article ">Figure 13
<p>MRI clustering results in the area of the southern longitudinal dune fields. Clustering results (<b>d</b>), Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>).</p>
Full article ">Figure 14
<p>MRI clustering results in the area with fluvial morphology, outcropping bedrock, and dome dune fields. Clustering results (<b>d</b>), Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>).</p>
Full article ">Figure 15
<p>Manual classification of crest lines (<b>a</b>) for large dunes using visual analysis of slope (<b>b</b>), profile curvature (<b>c</b>), and residual DEM (<b>d</b>). Crest lines are associated with high positive profile curvature, strongly positive residual DEM, and low slope. These locations are then located in areas in which the neighborhood is characterized by an abrupt variation in the selected geomorphometric derivatives.</p>
Full article ">Figure 16
<p>Probability of observing a crest obtained by means of RF considering the GDs integrated with the MRIs (<b>a</b>) and only the five GDs (<b>b</b>) to obtain details of the study area, which is located on the western mountain ridge. The RF model integrating the MRIs provides a more focused prediction of crest lines of large dunes. In (<b>c</b>), the prediction of the crest lines of the two RF models is compared. Pixels with a probability higher than 0.8 have been classified as crests. The transparent color is where both models predicted a not-crest pixel, green is where both models predicted a crest, and red and blue are where, respectively, only RF GDs and RF GDs + MRIs predicted a crest.</p>
Full article ">Figure 17
<p>Variables’ importance in the two RF models according to the mean decrease in the Gini index ((<b>a</b>), RF based on GDs; (<b>b</b>), RF based on GDs integrated with MRIs).</p>
Full article ">Figure 18
<p>Prediction of crest lines with the RF model based on GDs and MRIs of an unseen area ((<b>c</b>), green box) external to the one with reference data used for training and testing ((<b>c</b>), red box). The reference crest lines (<b>a</b>) have been manually digitized by means of visual analysis of the profile curvature, the residual DEM, and the slope; the predicted crest lines have been derived as crests of all of the pixels with a probability above 0.8. The predicted crest lines are compared with the reference data (<b>b</b>). Green pixels are correctly classified as crests; red and blue pixels are incorrectly classified, respectively, as crests and not crests.</p>
Full article ">
27 pages, 6983 KiB  
Article
DA-YOLOv7: A Deep Learning-Driven High-Performance Underwater Sonar Image Target Recognition Model
by Zhe Chen, Guohao Xie, Xiaofang Deng, Jie Peng and Hongbing Qiu
J. Mar. Sci. Eng. 2024, 12(9), 1606; https://doi.org/10.3390/jmse12091606 - 10 Sep 2024
Viewed by 1344
Abstract
Affected by the complex underwater environment and the limitations of low-resolution sonar image data and small sample sizes, traditional image recognition algorithms have difficulties achieving accurate sonar image recognition. The research builds on YOLOv7 and devises an innovative fast recognition model designed explicitly [...] Read more.
Affected by the complex underwater environment and the limitations of low-resolution sonar image data and small sample sizes, traditional image recognition algorithms have difficulties achieving accurate sonar image recognition. The research builds on YOLOv7 and devises an innovative fast recognition model designed explicitly for sonar images, namely the Dual Attention Mechanism YOLOv7 model (DA-YOLOv7), to tackle such challenges. New modules such as the Omni-Directional Convolution Channel Prior Convolutional Attention Efficient Layer Aggregation Network (OA-ELAN), Spatial Pyramid Pooling Channel Shuffling and Pixel-level Convolution Bilat-eral-branch Transformer (SPPCSPCBiFormer), and Ghost-Shuffle Convolution Enhanced Layer Aggregation Network-High performance (G-ELAN-H) are central to its design, which reduce the computational burden and enhance the accuracy in detecting small targets and capturing local features and crucial information. The study adopts transfer learning to deal with the lack of sonar image samples. By pre-training the large-scale Underwater Acoustic Target Detection Dataset (UATD dataset), DA-YOLOV7 obtains initial weights, fine-tuned on the smaller Smaller Common Sonar Target Detection Dataset (SCTD dataset), thereby reducing the risk of overfitting which is commonly encountered in small datasets. The experimental results on the UATD, the Underwater Optical Target Detection Intelligent Algorithm Competition 2021 Dataset (URPC), and SCTD datasets show that DA-YOLOV7 exhibits outstanding performance, with [email protected] scores reaching 89.4%, 89.9%, and 99.15%, respectively. In addition, the model maintains real-time speed while having superior accuracy and recall rates compared to existing mainstream target recognition models. These findings establish the superiority of DA-YOLOV7 in sonar image analysis tasks. Full article
Show Figures

Figure 1

Figure 1
<p>Structure of the YOLOv7.</p>
Full article ">Figure 2
<p>(<b>left</b>): OA-ELAN structure diagram, (<b>right</b>): ODConv Structure diagram.</p>
Full article ">Figure 3
<p>The CPCA attention mechanism.</p>
Full article ">Figure 4
<p><b>Left</b>: SPPCSPC structure, <b>Right</b>: BiFormer structure.</p>
Full article ">Figure 5
<p>Structure diagram of G-ELAN-H.</p>
Full article ">Figure 6
<p>The DA-YOLOv7 network.</p>
Full article ">Figure 7
<p>Confusion matrix of the ablation model: (<b>a</b>) YOLOv7; (<b>b</b>) YOLOv7 + OA-ELAN; (<b>c</b>) YOLOv7 + OA-ELAN + SPPCSPCBiFormer; (<b>d</b>) DA-YOLOv7.</p>
Full article ">Figure 8
<p>The PR curve: (<b>a</b>) YOLOv7; (<b>b</b>) YOLOv7 + OA-ELAN; (<b>c</b>) YOLOv7 + OA-ELAN + SPPCSPCBiFormer; (<b>d</b>) DA-YOLOv7.</p>
Full article ">Figure 9
<p>Curve of the change in loss value on the UATD dataset.</p>
Full article ">Figure 10
<p>Prediction results of various targets in UATD multi-beam forward-looking sonar images.</p>
Full article ">Figure 11
<p>SCTD sonar image dataset: (<b>a</b>) human; (<b>b</b>) ship; (<b>c</b>) aircraft.</p>
Full article ">Figure 12
<p>Flowchart of the training strategy for the SCTD dataset.</p>
Full article ">Figure 13
<p>The effect of recognition on the SCTD dataset: (<b>a</b>) SCTD mAP Results; (<b>b</b>) SCTD aircraft-AP results; (<b>c</b>) SCTD human-AP results; (<b>d</b>) SCTD ship-AP results.</p>
Full article ">Figure 14
<p>The sample information of URPC is as follows: (<b>a</b>) Labels: The upper left corner shows the distribution of categories; the upper right corner presents the visualization of all box sizes; the lower left corner indicates the distribution of the box centroid position; the lower right corner depicts the distribution of the box aspect ratio. (<b>b</b>) Example images.</p>
Full article ">Figure 15
<p>Recognition results in multiple underwater scenes.</p>
Full article ">Figure 16
<p>RCurve of loss value changes on the UPRC dataset.</p>
Full article ">
25 pages, 28511 KiB  
Article
A Method for Estimating the Distribution of Trachinotus ovatus in Marine Cages Based on Omnidirectional Scanning Sonar
by Yu Hu, Jiazhen Hu, Pengqi Sun, Guohao Zhu, Jialong Sun, Qiyou Tao, Taiping Yuan, Gen Li, Guoliang Pang and Xiaohua Huang
J. Mar. Sci. Eng. 2024, 12(9), 1571; https://doi.org/10.3390/jmse12091571 - 6 Sep 2024
Viewed by 572
Abstract
In order to accurately estimate the distribution of Trachinotus ovatus in marine cages, a novel method was developed using omnidirectional scanning sonar and deep-learning techniques. This method involved differentiating water layers and clustering data layer by layer to achieve precise location estimation. The [...] Read more.
In order to accurately estimate the distribution of Trachinotus ovatus in marine cages, a novel method was developed using omnidirectional scanning sonar and deep-learning techniques. This method involved differentiating water layers and clustering data layer by layer to achieve precise location estimation. The approach comprised two main components: fish identification and fish clustering. Firstly, omnidirectional scanning sonar was employed to perform spiral detection within marine cages, capturing fish image data. These images were then labeled to construct a training dataset for an enhanced CS-YOLOv8s model. After training, the CS-YOLOv8s model was used to identify and locate fish within the images. Secondly, the cages were divided into water layers with depth intervals of 40 cm. The identification coordinate data for each water layer were clustered using the DBSCAN method to generate location coordinates for the fish in each layer. Finally, the coordinate data from all water layers were consolidated to determine the overall distribution of fish within the cage. This method was shown, through multiple experimental results, to effectively estimate the distribution of Trachinotus ovatus in marine cages, closely matching the distributions detected manually. Full article
(This article belongs to the Special Issue New Techniques and Equipment in Large Offshore Aquaculture Platform)
Show Figures

Figure 1

Figure 1
<p>Satellite image map of the experimental site.</p>
Full article ">Figure 2
<p>Aerial view of the experimental site.</p>
Full article ">Figure 3
<p>Cage used in the experiment.</p>
Full article ">Figure 4
<p><span class="html-italic">Trachinotus ovatus</span> used in the experiment. (<b>a</b>) Body length; (<b>b</b>) Body height.</p>
Full article ">Figure 5
<p>Working principle diagram of omnidirectional scanning sonar. (<b>a</b>) Schematic diagram of sonar scanning; (<b>b</b>) Work diagram.</p>
Full article ">Figure 6
<p>Omnidirectional scanning sonar.</p>
Full article ">Figure 7
<p>Sonar Assembly.</p>
Full article ">Figure 8
<p>Side view of the sea cage. The sonar in the yellow box is located on the center axis of the net cage, and the yellow dotted lines shows the range covered by the sonar.</p>
Full article ">Figure 9
<p>Training process (Loss).</p>
Full article ">Figure 10
<p>Training process (mAP@0.5%).</p>
Full article ">Figure 11
<p>Coordinate transformation. (<b>a</b>) A coordinate system with sonar as the origin; (<b>b</b>) A coordinate system with the top left corner as the origin.</p>
Full article ">Figure 12
<p>Water layer division.</p>
Full article ">Figure 13
<p>Clustering in the 100–140 cm water layer. (<b>a</b>) Fish distribution; (<b>b</b>) Cluster effect diagram.</p>
Full article ">Figure 14
<p>Water layer data and the data after clustering. (<b>a</b>) Fish distribution map before clustering; (<b>b</b>) Cluster diagram of noisy points.</p>
Full article ">Figure 15
<p>Center and noise point of core object. (<b>a</b>) Distribution map of center point and noise point; (<b>b</b>) Centralization rendering.</p>
Full article ">Figure 16
<p>Sonar images of different numbers of fish. (<b>a</b>) 100 fish; (<b>b</b>) 150 fish; (<b>c</b>) 200 fish.</p>
Full article ">Figure 17
<p>Fish for object identification. (<b>a</b>) 100 fish; (<b>b</b>) 150 fish; (<b>c</b>) 200 fish.</p>
Full article ">Figure 18
<p>Trend of number of fish with depth. (<b>a</b>) Comparison bar chart of depth distribution; (<b>b</b>) Comparison line chart of fish depth distribution.</p>
Full article ">Figure 19
<p>Horizontal distribution maps of fish. (<b>a</b>) Horizontal distribution of 100 fish; (<b>b</b>) Horizontal distribution of 150 fish; (<b>c</b>) Horizontal distribution of 200 fish.</p>
Full article ">Figure 20
<p>Horizontal distribution comparison chart of fish.</p>
Full article ">Figure 21
<p>Spatial distribution maps of 100 Fish in each water layer. (<b>a</b>) Water layer: 0–40 cm; (<b>b</b>) Water layer: 25–65 cm; (<b>c</b>) Water layer: 50–90 cm; (<b>d</b>) Water layer: 75–115 cm; (<b>e</b>) Water layer: 100–140 cm; (<b>f</b>) Water layer: 125–165 cm; (<b>g</b>) Water layer: 150–190 cm; (<b>h</b>) Water layer: 175–215 cm; (<b>i</b>) Water layer: 200–240 cm; (<b>j</b>) Water layer: 225–265 cm; (<b>k</b>) Water layer: 250–290 cm.</p>
Full article ">Figure 22
<p>Spatial distribution maps of 150 Fish in each water layer. (<b>a</b>) Water layer: 0–40 cm; (<b>b</b>) Water layer: 25–65 cm; (<b>c</b>) Water layer: 50–90 cm; (<b>d</b>) Water layer: 75–115 cm; (<b>e</b>) Water layer: 100–140 cm; (<b>f</b>) Water layer: 125–165 cm; (<b>g</b>) Water layer: 150–190 cm; (<b>h</b>) Water layer: 175–215 cm; (<b>i</b>) Water layer: 200–240 cm; (<b>j</b>) Water layer: 225–265 cm; (<b>k</b>) Water layer: 250–290 cm.</p>
Full article ">Figure 23
<p>Spatial distribution maps of 200 fish in each water layer. (<b>a</b>) Water layer: 0–40 cm; (<b>b</b>) Water layer: 25–65 cm; (<b>c</b>) Water layer: 50–90 cm; (<b>d</b>) Water layer: 75–115 cm; (<b>e</b>) Water layer: 100–140 cm; (<b>f</b>) Water layer: 125–165 cm; (<b>g</b>) Water layer: 150–190 cm; (<b>h</b>) Water layer: 175–215 cm; (<b>i</b>) Water layer: 200–240 cm; (<b>j</b>) Water layer: 225–265 cm; (<b>k</b>) Water layer: 250–290 cm.</p>
Full article ">Figure 24
<p>3D spatial distribution maps of cage fish. (<b>a</b>) Spatial distribution of 100 fish group; (<b>b</b>) Spatial distribution of 150 fish group; (<b>c</b>) Spatial distribution of 200 fish group.</p>
Full article ">Figure 25
<p>Spatial distribution comparison chart of fish.</p>
Full article ">Figure 26
<p>Distribution of fish at different temperatures. (<b>a</b>) Distribution of 100 fish; (<b>b</b>) Distribution of 150 fish; (<b>c</b>) Distribution of 200 fish.</p>
Full article ">
13 pages, 3604 KiB  
Article
A Super-Resolution and 3D Reconstruction Method Based on OmDF Endoscopic Images
by Fujia Sun and Wenxuan Song
Sensors 2024, 24(15), 4890; https://doi.org/10.3390/s24154890 - 27 Jul 2024
Viewed by 1298
Abstract
In the field of endoscopic imaging, challenges such as low resolution, complex textures, and blurred edges often degrade the quality of 3D reconstructed models. To address these issues, this study introduces an innovative endoscopic image super-resolution and 3D reconstruction technique named Omni-Directional Focus [...] Read more.
In the field of endoscopic imaging, challenges such as low resolution, complex textures, and blurred edges often degrade the quality of 3D reconstructed models. To address these issues, this study introduces an innovative endoscopic image super-resolution and 3D reconstruction technique named Omni-Directional Focus and Scale Resolution (OmDF-SR). This method integrates an Omnidirectional Self-Attention (OSA) mechanism, an Omnidirectional Scale Aggregation Group (OSAG), a Dual-stream Adaptive Focus Mechanism (DAFM), and a Dynamic Edge Adjustment Framework (DEAF) to enhance the accuracy and efficiency of super-resolution processing. Additionally, it employs Structure from Motion (SfM) and Multi-View Stereo (MVS) technologies to achieve high-precision medical 3D models. Experimental results indicate significant improvements in image processing with a PSNR of 38.2902 dB and an SSIM of 0.9746 at a magnification factor of ×2, and a PSNR of 32.1723 dB and an SSIM of 0.9489 at ×4. Furthermore, the method excels in reconstructing detailed 3D models, enhancing point cloud density, mesh quality, and texture mapping richness, thus providing substantial support for clinical diagnosis and surgical planning. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>The structure of Omni Self-Attention (OSA).</p>
Full article ">Figure 2
<p>The structure of Local Convolution Block (LCB).</p>
Full article ">Figure 3
<p>The structure of the Dual-stream Adaptive Focus Mechanism (DAFM).</p>
Full article ">Figure 4
<p>The overall architecture of the proposed OmDF-SR framework.</p>
Full article ">Figure 5
<p>The overall architecture of 3D reconstruction.</p>
Full article ">Figure 6
<p>Training loss of different model.</p>
Full article ">Figure 7
<p>Perceptual results of different models, corresponding HR image, and initial image with enlargement scale factor ×2.</p>
Full article ">Figure 8
<p>Reconstructed 3D model.</p>
Full article ">
22 pages, 7448 KiB  
Article
End-to-End Framework for the Automatic Matching of Omnidirectional Street Images and Building Data and the Creation of 3D Building Models
by Yoshiki Ogawa, Ryoto Nakamura, Go Sato, Hiroya Maeda and Yoshihide Sekimoto
Remote Sens. 2024, 16(11), 1858; https://doi.org/10.3390/rs16111858 - 23 May 2024
Viewed by 1055
Abstract
For accurate urban planning, three-dimensional (3D) building models with a high level of detail (LOD) must be developed. However, most large-scale 3D building models are limited to a low LOD of 1–2, as the creation of higher LOD models requires the modeling of [...] Read more.
For accurate urban planning, three-dimensional (3D) building models with a high level of detail (LOD) must be developed. However, most large-scale 3D building models are limited to a low LOD of 1–2, as the creation of higher LOD models requires the modeling of detailed building elements such as walls, windows, doors, and roof shapes. This process is currently not automated and is performed manually. In this study, an end-to-end framework for the creation of 3D building models was proposed by integrating multi-source data such as omnidirectional images, building footprints, and aerial photographs. These different data sources were matched with the building ID considering their spatial location. The building element information related to the exterior of the building was extracted, and detailed LOD3 3D building models were created. Experiments were conducted using data from Kobe, Japan, yielding a high accuracy for the intermediate processes, such as an 86.9% accuracy in building matching, an 88.3% pixel-based accuracy in the building element extraction, and an 89.7% accuracy in the roof type classification. Eighty-one LOD3 3D building models were created in 8 h, demonstrating that our method can create 3D building models that adequately represent the exterior information of actual buildings. Full article
Show Figures

Figure 1

Figure 1
<p>Five levels of detail in CityGML [Reprinted/adapted with permission from Ref. [<a href="#B9-remotesensing-16-01858" class="html-bibr">9</a>]. 2022, MLIT Japan].</p>
Full article ">Figure 2
<p>Building data used for this study (shooting points of omnidirectional images, building footprints, and aerial photographs). In our study, approximately, 100 m × 100 m of the orange-colored buildings were reconstructed.</p>
Full article ">Figure 3
<p>Examples of texture images.</p>
Full article ">Figure 4
<p>Overview of the proposed framework.</p>
Full article ">Figure 5
<p>The flow of matching omnidirectional images and buildings.</p>
Full article ">Figure 6
<p>Schematic representation of the texture synthesis model.</p>
Full article ">Figure 7
<p>Types of building roofs.</p>
Full article ">Figure 8
<p>Examples of images linked to edges of a building polygon with ID 314960.</p>
Full article ">Figure 9
<p>Results of the semantic segmentation: (<b>a</b>) original building image, (<b>b</b>) ground truth, and (<b>c</b>) prediction.</p>
Full article ">Figure 10
<p>Results of window and door location detection: (<b>a</b>) original building image and (<b>b</b>) prediction.</p>
Full article ">Figure 11
<p>Results of the texture synthesis: (<b>a</b>) original building image, (<b>b</b>) wall crop image, and (<b>c</b>) synthesized texture image.</p>
Full article ">Figure 12
<p>Results of roof color extraction: (<b>a</b>) building polygon cut from an aerial photograph, (<b>b</b>) average color of the polygon, and (<b>c</b>) extracted color using the proposed method.</p>
Full article ">Figure 13
<p>Comparison of the created 3D model with the Google Earth model of the target area.</p>
Full article ">Figure 14
<p>Individual examples of generated 3D models: (<b>a</b>) created 3D model and (<b>b</b>) Google Earth model.</p>
Full article ">
16 pages, 4187 KiB  
Article
An Omnidirectional Image Super-Resolution Method Based on Enhanced SwinIR
by Xiang Yao, Yun Pan and Jingtao Wang
Information 2024, 15(5), 248; https://doi.org/10.3390/info15050248 - 28 Apr 2024
Cited by 1 | Viewed by 1412
Abstract
For the significant distortion problem caused by the special projection method of equi-rectangular projection (ERP) images, this paper proposes an omnidirectional image super-resolution algorithm model based on position information transformation, taking SwinIR as the base. By introducing a space position transformation module that [...] Read more.
For the significant distortion problem caused by the special projection method of equi-rectangular projection (ERP) images, this paper proposes an omnidirectional image super-resolution algorithm model based on position information transformation, taking SwinIR as the base. By introducing a space position transformation module that supports deformable convolution, the image preprocessing process is optimized to reduce the distortion effects in the polar regions of the ERP image. Meanwhile, by introducing deformable convolution in the deep feature extraction process, the model’s adaptability to local deformations of images is enhanced. Experimental results on publicly available datasets have shown that our method outperforms SwinIR, with an average improvement of over 0.2 dB in WS-PSNR and over 0.030 in WS-SSIM for ×4 pixel upscaling. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>The Network Architecture of Our Proposed Model (This model aims to convert low resolution (LR) images into high resolution (HR) images).</p>
Full article ">Figure 2
<p>LTM Module.</p>
Full article ">Figure 3
<p>Affine Transformation Diagram.</p>
Full article ">Figure 4
<p>The Process of Deformable Convolution.</p>
Full article ">Figure 5
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 6
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 7
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 8
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 9
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 10
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 11
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 12
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 13
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 14
<p>Changes in Results as Training Progresses.</p>
Full article ">Figure 15
<p>Changes in Loss Function.</p>
Full article ">
22 pages, 7518 KiB  
Article
Omni-OTPE: Omnidirectional Optimal Real-Time Ground Target Position Estimation System for Moving Lightweight Unmanned Aerial Vehicle
by Yi Ding, Jiaxing Che, Zhiming Zhou and Jingyuan Bian
Sensors 2024, 24(5), 1709; https://doi.org/10.3390/s24051709 - 6 Mar 2024
Viewed by 1181
Abstract
Ground target detection and positioning systems based on lightweight unmanned aerial vehicles (UAVs) are increasing in value for aerial reconnaissance and surveillance. However, the current method for estimating the target’s position is limited by the field of view angle, rendering it challenging to [...] Read more.
Ground target detection and positioning systems based on lightweight unmanned aerial vehicles (UAVs) are increasing in value for aerial reconnaissance and surveillance. However, the current method for estimating the target’s position is limited by the field of view angle, rendering it challenging to fulfill the demands of a real-time omnidirectional reconnaissance operation. To address this issue, we propose an Omnidirectional Optimal Real-Time Ground Target Position Estimation System (Omni-OTPE) that utilizes a fisheye camera and LiDAR sensors. The object of interest is first identified in the fisheye image, and then, the image-based target position is obtained by solving using the fisheye projection model and the target center extraction algorithm based on the detected edge information. Next, the LiDAR’s real-time point cloud data are filtered based on position–direction constraints using the image-based target position information. This step allows for the determination of point cloud clusters that are relevant to the characterization of the target’s position information. Finally, the target positions obtained from the two methods are fused using an optimal Kalman fuser to obtain the optimal target position information. In order to evaluate the positioning accuracy, we designed a hardware and software setup, mounted on a lightweight UAV, and tested it in a real scenario. The experimental results validate that our method exhibits significant advantages over traditional methods and achieves a real-time high-performance ground target position estimation function. Full article
Show Figures

Figure 1

Figure 1
<p>Omni-OTPE: Omnidirectional optimal real-time ground target position estimation system framework.</p>
Full article ">Figure 2
<p>The drone is equipped with Omni-OTPE system that allows it to estimate the position of a target-tagged object in conjunction with its own position.</p>
Full article ">Figure 3
<p>The layout of the sensors and the range of vertical sensing angles.</p>
Full article ">Figure 4
<p>Yolov5 performance plots for each model.</p>
Full article ">Figure 5
<p>Comparison of distortion of different projection models. The model includes Keyhole projection, Stereo graphic projection, Equidistance projection, Equisolid projection and Orthogonal projection. Pinhole projection is the standard camera model, which theoretically has no distortion. The degree of distortion of the model can be expressed as the degree of curve deviation between this model and pinhole projection model.</p>
Full article ">Figure 6
<p>Description of the center offset problem.</p>
Full article ">Figure 7
<p>(<b>a</b>) Polar and azimuthal angles corresponding to the upper left (red) and lower right (green) fixed points of the identification box in the fisheye image. (<b>b</b>) Fisheye camera point projection process.</p>
Full article ">Figure 8
<p>(<b>a</b>) Position-based point cloud filtering. (<b>b</b>) Direction-based point cloud filtering, where the green dots are the point cloud generated by the target; the red dots indicate anomalies; and the blue dots indicate vision-based position estimates <math display="inline"><semantics> <mrow> <mmultiscripts> <mi>P</mi> <mi mathvariant="italic">tag</mi> <mi>V</mi> <mprescripts/> <none/> <mi>w</mi> </mmultiscripts> </mrow> </semantics></math>.</p>
Full article ">Figure 9
<p>Time series of the target location output.</p>
Full article ">Figure 10
<p>Optimal target position information fusion process.</p>
Full article ">Figure 11
<p>Scene detection graph. There are three box objects of similar size in the scene. The one with the gray surface and label is the target.</p>
Full article ">Figure 12
<p>(<b>a</b>) Flight experiment in which there are two distractors and one target object; (<b>b</b>) UAV trajectory and target position estimates, where the color of the trajectory can indicate the distance from the UAV to the target.</p>
Full article ">Figure 13
<p>Image-based plot of horizontal coordinates of the estimated points of the target position, indicating the distribution of the estimated points. (<b>a</b>) Estimated target position at 5 m altitude by the UAV. where the ellipse denotes the 95 confidence ellipse, which can indicate the degree of distribution of the position estimates. (<b>b</b>) Estimated target position at 7 m altitude by the UAV.</p>
Full article ">Figure 14
<p>Direction angle error curves for both methods. (<b>a</b>) Angular error in position estimation in flight at 5 m altitude. (<b>b</b>) Angular error in position estimation in flight at 7 m altitude.</p>
Full article ">Figure 15
<p>Target position estimation error curves for both methods. (<b>a</b>) Positioning error in position estimation in flight at 5 m altitude. (<b>b</b>) Positioning error in position estimation in flight at 7 m altitude.</p>
Full article ">Figure 16
<p>When the visual localization error is too large, the back-end will reject the correct point cloud information, reducing the back-end repositioning performance.</p>
Full article ">Figure 17
<p>Three positional estimation error curves. (<b>a</b>) Target localization error of each localization method in 5 m flight experiment. (<b>b</b>) Target localization error of each localization method in 7 m flight experiment. The red point is the error of LiDAR localization, and the red area indicates that the system finds the corresponding target point cloud cluster at that moment based on the visual localization result. The blue curve is the position error curve after fusion.</p>
Full article ">
21 pages, 1865 KiB  
Article
Latitude-Adaptive Integer Bit Allocation for Quantization of Omnidirectional Images
by Qian Sima, Hui Feng and Bo Hu
Appl. Sci. 2024, 14(5), 1861; https://doi.org/10.3390/app14051861 - 23 Feb 2024
Cited by 1 | Viewed by 1002
Abstract
Omnidirectional images have gained significant popularity and drawn great attention nowadays, which poses challenges to omnidirectional image processing in solving the bottleneck of storage and transmission. Projecting onto a two-dimensional image plane is generally used to compress an omnidirectional image. However, the most [...] Read more.
Omnidirectional images have gained significant popularity and drawn great attention nowadays, which poses challenges to omnidirectional image processing in solving the bottleneck of storage and transmission. Projecting onto a two-dimensional image plane is generally used to compress an omnidirectional image. However, the most commonly used projection format, the equirectangular projection (ERP), results in a significant amount of redundant samples in the polar areas, thus incurring extra bitrate and geometric distortion. We derive the optimal latitude-adaptive bit allocation for each image tile. Subsequently, we propose a greedy algorithm for non-negative integer bit allocation (NNIBA) for non-uniform quantization under an omnidirectional image quality metric WMSE. In our experiment, we design quantization tables based on JPEG and compare our approach with other sampling-related methods. Our method achieves an average bit saving of 7.9% compared with JPEG while outperforming other sampling-related methods. Besides, we compare our non-uniform quantization approach with two proposed bit allocation methods, achieving an average improvement of 0.35 dB and 2.66 dB under WS-PSNR, respectively. The visual quality assessment also confirms the superiority of our method. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of omnidirectional image compression methods.</p>
Full article ">Figure 2
<p>Processing chain of our method. (<b>a</b>) A batch of original images. (<b>b</b>) Coefficient matrices partitioned into blocks. (<b>c</b>) Recovered images after quantization. (<b>d</b>) Distribution modeling of coefficients at different latitudes. (<b>e</b>) Bit allocation among titles. (<b>f</b>) Subsequent bit allocation in each tile.</p>
Full article ">Figure 3
<p>Coefficient frequencies at latitude 1. (<b>a</b>) DC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>1</mn> </mrow> </msub> </semantics></math>. (<b>b</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> </msub> </semantics></math>. (<b>c</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>30</mn> </mrow> </msub> </semantics></math>. (<b>d</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>64</mn> </mrow> </msub> </semantics></math>.</p>
Full article ">Figure 4
<p>Coefficients frequencies at latitude 17. (<b>a</b>) DC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>17</mn> <mo>,</mo> <mn>1</mn> </mrow> </msub> </semantics></math>. (<b>b</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>17</mn> <mo>,</mo> <mn>2</mn> </mrow> </msub> </semantics></math>. (<b>c</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>17</mn> <mo>,</mo> <mn>30</mn> </mrow> </msub> </semantics></math>. (<b>d</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>17</mn> <mo>,</mo> <mn>64</mn> </mrow> </msub> </semantics></math>.</p>
Full article ">Figure 5
<p>Rate distortion curves for JPEG-based methods over 100 images with a resolution of <math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>1024</mn> </mrow> </semantics></math> pixels [<a href="#B10-applsci-14-01861" class="html-bibr">10</a>,<a href="#B20-applsci-14-01861" class="html-bibr">20</a>,<a href="#B23-applsci-14-01861" class="html-bibr">23</a>].</p>
Full article ">Figure 6
<p>Rate distortion curves for non-negative integer bit allocation (NNIBA) methods over 100 images with a resolution of <math display="inline"><semantics> <mrow> <mn>256</mn> <mo>×</mo> <mn>512</mn> </mrow> </semantics></math> pixels [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 7
<p>Rate distortion curves for non-negative integer bit allocation (NNIBA) methods over 100 images with a resolution of <math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>1024</mn> </mrow> </semantics></math> pixels [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 8
<p>Four test images from Pano 3D dataset with a resolution of <math display="inline"><semantics> <mrow> <mn>256</mn> <mo>×</mo> <mn>512</mn> </mrow> </semantics></math>. The red box is used for subsequent zoomed comparisons.</p>
Full article ">Figure 9
<p>Zoomed version of test image A at 0.4 bpp [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 10
<p>Zoomed version of test image B at 0.4 bpp [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 11
<p>Zoomedversion of test image C at 0.75 bpp [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 12
<p>Zoomedversion of test image D at 0.75 bpp [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">
18 pages, 10954 KiB  
Article
Using a Robot for Indoor Navigation and Door Opening Control Based on Image Processing
by Chun-Hsiang Hsu and Jih-Gau Juang
Actuators 2024, 13(2), 78; https://doi.org/10.3390/act13020078 - 16 Feb 2024
Viewed by 1705
Abstract
This study used real-time image processing to realize obstacle avoidance and indoor navigation with an omnidirectional wheeled mobile robot (WMR). The distance between an obstacle and the WMR was obtained using a depth camera. Real-time images were used to control the robot’s movements. [...] Read more.
This study used real-time image processing to realize obstacle avoidance and indoor navigation with an omnidirectional wheeled mobile robot (WMR). The distance between an obstacle and the WMR was obtained using a depth camera. Real-time images were used to control the robot’s movements. The WMR can extract obstacle distance data from a depth map and apply fuzzy theory to avoid obstacles in indoor environments. A fuzzy control system was integrated into the control scheme. After detecting a doorknob, the robot could track the target and open the door. We used the speeded up robust features matching algorithm to recognize the WMR’s movement direction. The proposed control scheme ensures that the WMR can avoid obstacles, move to a designated location, and open a door. Like humans, the robot performs the described task only using visual sensors. Full article
(This article belongs to the Special Issue Actuators in Robotic Control—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>The primary devices of the proposed robot system are (<b>a</b>) a mobile robot; (<b>b</b>) an omnidirectional wheel, battery, and control components; and (<b>c</b>) an Intel Realsense depth camera.</p>
Full article ">Figure 2
<p>(<b>a</b>) Omnidirectional wheel structure. (<b>b</b>) Coordinate system.</p>
Full article ">Figure 3
<p>(<b>a</b>) Arduino Uno R3 [<a href="#B15-actuators-13-00078" class="html-bibr">15</a>], (<b>b</b>) DFRduino IO Expansion board, and (<b>c</b>) Microsoft LifeCam.</p>
Full article ">Figure 4
<p>Zhang Zhengyou camera calibration samples.</p>
Full article ">Figure 5
<p>Stereo rectification.</p>
Full article ">Figure 6
<p>Mug distance.</p>
Full article ">Figure 7
<p>(<b>a</b>) is the original image, (<b>b</b>) is the depth image where the blue color means the object is near to the camera, and dark red means the object is far from the camera. The distance ranges from 0 m (dark blue) to 8 m (dark red).</p>
Full article ">Figure 8
<p>Detection of different obstacles: (<b>a</b>) obstacle detection and obstacle depth map, (<b>b</b>) nearest obstacle point is on the right side, (<b>c</b>) move to the left and the obstacle is outside the safe frame, (<b>d</b>) nearest obstacle point is on the left side, (<b>e</b>) move to the right and the moving direction is clear.</p>
Full article ">Figure 9
<p>SURF matching: (<b>a</b>) robot is heading in the right direction, (<b>b</b>) robot is heading in the wrong direction, (<b>c</b>) robot is heading in the right direction.</p>
Full article ">Figure 10
<p>Edge detection uses a threshold, and the doorknob is detected; the Chinese character on the left part of the figure is the room’s name.</p>
Full article ">Figure 11
<p>Control flowchart.</p>
Full article ">Figure 12
<p>Robot moving path.</p>
Full article ">Figure 13
<p>The fuzzy sets of the three inputs are near (blue), medium (red), and far (yellow); the fuzzy sets of the output pixel are turn_left (blue), go_straight (red), and turn_right (yellow); the fuzzy sets of the output time are short (blue), medium (red), and long (yellow).</p>
Full article ">Figure 14
<p>Fuzzy control model where the yellow color means large value and the dark blue means small value.</p>
Full article ">Figure 15
<p>Fuzzy control scheme.</p>
Full article ">Figure 16
<p>The robot arm lifts to the specified height; the Chinese characters on the figure are the room’s name.</p>
Full article ">Figure 17
<p>Robot obstacle avoidance test.</p>
Full article ">Figure 18
<p>On the depth camera image, the distances of the nearest left, middle, and right objects are greater than 2 m, so the action is GO.</p>
Full article ">Figure 19
<p>Robot’s starting position.</p>
Full article ">Figure 20
<p>The robot finds an obstacle as the blue mark (in the middle area) and yellow mark (in the right area) on the picture, and the distances are 1.65 m and 1.62 m, respectively; the action is a LEFT turn.</p>
Full article ">Figure 21
<p>Robot avoids obstacle (box).</p>
Full article ">Figure 22
<p>Image on the robot’s depth camera.</p>
Full article ">Figure 23
<p>(<b>a</b>) The robot reaches the midway point; (<b>b</b>) the robot’s image.</p>
Full article ">Figure 24
<p>After the SURF matching, the robot turns to the target direction and moves forward.</p>
Full article ">Figure 25
<p>(<b>a</b>) The robot arrives at the specified location; (<b>b</b>) the Chinese characters on the figure are the room’s name.</p>
Full article ">Figure 26
<p>(<b>a</b>) Doorknob detection, Activate arm and arm’s camera; (<b>b</b>) the Chinese characters on the figure are the room’s name. Arm’s camera image; the object distance is 0.827 m, and the coordinate is (436.5, 51.5).</p>
Full article ">Figure 27
<p>The robot opens the door; the Chinese characters on the figure are the room’s name.</p>
Full article ">
16 pages, 9434 KiB  
Article
Omnidirectional-Sensor-System-Based Texture Noise Correction in Large-Scale 3D Reconstruction
by Wenya Xie and Xiaoping Hong
Sensors 2024, 24(1), 78; https://doi.org/10.3390/s24010078 - 22 Dec 2023
Viewed by 1117
Abstract
The evolution of cameras and LiDAR has propelled the techniques and applications of three-dimensional (3D) reconstruction. However, due to inherent sensor limitations and environmental interference, the reconstruction process often entails significant texture noise, such as specular highlight, color inconsistency, and object occlusion. Traditional [...] Read more.
The evolution of cameras and LiDAR has propelled the techniques and applications of three-dimensional (3D) reconstruction. However, due to inherent sensor limitations and environmental interference, the reconstruction process often entails significant texture noise, such as specular highlight, color inconsistency, and object occlusion. Traditional methodologies grapple to mitigate such noise, particularly in large-scale scenes, due to the voluminous data produced by imaging sensors. In response, this paper introduces an omnidirectional-sensor-system-based texture noise correction framework for large-scale scenes, which consists of three parts. Initially, we obtain a colored point cloud with luminance value through LiDAR points and RGB images organization. Next, we apply a voxel hashing algorithm during the geometry reconstruction to accelerate the computation speed and save the computer memory. Finally, we propose the key innovation of our paper, the frame-voting rendering and the neighbor-aided rendering mechanisms, which effectively eliminates the aforementioned texture noise. From the experimental results, the processing rate of one million points per second shows its real-time applicability, and the output figures of texture optimization exhibit a significant reduction in texture noise. These results indicate that our framework has advanced performance in correcting multiple texture noise in large-scale 3D reconstruction. Full article
(This article belongs to the Special Issue Sensing and Processing for 3D Computer Vision: 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Specular highlight phenomenon. (<b>b</b>) The position of the highlight areas in the image changes with the variation of the sensor pose. In the image, the red box indicates the most prominent highlight noise, and the green box indicates the door, which serves as a positional reference.</p>
Full article ">Figure 2
<p>Color inconsistency phenomenon. P1–P3 are three consecutive images in terms of position. (<b>a</b>) Normal situation with consistent color between frames. (<b>b</b>) Inconsistent color between frames caused by variations in the intensity of the light source or changes in its relative position to the sensor.</p>
Full article ">Figure 3
<p>Pipeline of the whole process, consisting of data organization, geometry reconstruction, and texture optimization.</p>
Full article ">Figure 4
<p>Process flow of data organization. (<b>a</b>) RGB image. (<b>b</b>) CIELAB color space image transformed from RGB image, which facilitates luminance evaluation in the subsequent section of our work. (<b>c</b>) LiDAR point cloud. (<b>d</b>) Fusion of LiDAR point cloud with RGB image. (<b>e</b>) Fusion of LiDAR point cloud with CIELAB color space image.</p>
Full article ">Figure 5
<p>Voxel hashing schematic. The mapping between point coordinates and voxel block indices is achieved through a hash table, thereby efficiently allocating points while making reasonable use of computer storage resources.</p>
Full article ">Figure 6
<p>Motivation for proposing neighbor-aided rendering mechanism: points are randomly distributed in voxels; thus, some voxels lack insufficient points for self-optimization.</p>
Full article ">Figure 7
<p>Neighbor-aided rendering mechanism. The figure illustrates the configuration of a voxel block and the interconnections between adjacent voxels.</p>
Full article ">Figure 8
<p>Sensor setup for data collection.</p>
Full article ">Figure 9
<p>Input data. The dataset consists of four spots, and each spot consists of five specified poses.</p>
Full article ">Figure 10
<p>Highlight noise correction in scene 1 according to frame-voting rendering. Regions (<b>a</b>)–(<b>c</b>) present specular highlights phenomenon on the screen and wall surfaces in the scene.</p>
Full article ">Figure 11
<p>Elimination of object occlusion in scene 2 with frame-voting rendering. (<b>a</b>) Comparison diagram of the elimination of misimaging caused by table occlusion. (<b>b</b>) Comparison diagram of the elimination of misimaging caused by chair occlusion.</p>
Full article ">Figure 12
<p>Enhanced outcome with neighbor-aided optimization. Regions A–C exhibite pronounced contrastive effects. (<b>a</b>) Demonstration area of the original point cloud containing numerous types of texture noise. (<b>b</b>) The result optimized using only frame-voting rendering. (<b>c</b>) The result optimized further with neighbor-aided rendering.</p>
Full article ">Figure 13
<p>Comparing results of highlight removal method. (<b>a</b>) Projection of raw model (input). The white boxes indicate areas with noise that should be corrected. The red box indicates area that should not be corrected (lights). (<b>b</b>) Projection of texture optimized model (ours). (<b>c</b>) Yang et al. (2010) [<a href="#B2-sensors-24-00078" class="html-bibr">2</a>]. (<b>d</b>) Shen et al. (2013) [<a href="#B3-sensors-24-00078" class="html-bibr">3</a>]. (<b>e</b>) Fu et al. (2019) [<a href="#B4-sensors-24-00078" class="html-bibr">4</a>]. (<b>f</b>) Jin et al. (2023) [<a href="#B8-sensors-24-00078" class="html-bibr">8</a>].</p>
Full article ">
21 pages, 54036 KiB  
Article
Plant Foliar Geometry as a Biomimetic Template for Antenna Design
by Jose Ignacio Lozano, Marco A. Panduro, Rodrigo Méndez-Alonzo, Miguel A. Alonso-Arevalo, Roberto Conte and Alberto Reyna
Biomimetics 2023, 8(7), 531; https://doi.org/10.3390/biomimetics8070531 - 7 Nov 2023
Cited by 1 | Viewed by 1826
Abstract
Plant diversity includes over 300,000 species, and leaf structure is one of the main targets of selection, being highly variable in shape and size. On the other hand, the optimization of antenna design has no unique solution to satisfy the current range of [...] Read more.
Plant diversity includes over 300,000 species, and leaf structure is one of the main targets of selection, being highly variable in shape and size. On the other hand, the optimization of antenna design has no unique solution to satisfy the current range of applications. We analyzed the foliar geometries of 100 plant species and applied them as a biomimetic design template for microstrip patch antenna systems. From this set, a subset of seven species were further analyzed, including species from tropical and temperate forests across the phylogeny of the Angiosperms. Foliar geometry per species was processed by image processing analyses, and the resultant geometries were used in simulations of the reflection coefficients and the radiation patterns via finite differences methods. A value below −10 dB is set for the reflection coefficient to determine the operation frequencies of all antenna elements. All species showed between 3 and 15 operational frequencies, and four species had operational frequencies that included the 2.4 and 5 GHz bands. The reflection coefficients and the radiation patterns in most of the designs were equal or superior to those of conventional antennas, with several species showing multiband effects and omnidirectional radiation. We demonstrate that plant structures can be used as a biomimetic tool in designing microstrip antenna for a wide range of applications. Full article
(This article belongs to the Special Issue Bio-Inspired Design: Creativity and Innovation)
Show Figures

Figure 1

Figure 1
<p>Basic method to assess the structure or each leaf design as a radiating element.</p>
Full article ">Figure 2
<p><span class="html-italic">Tetrapterys macrocarpa</span> Malpighiaceae (elliptic) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> parameter, (<b>c</b>) 3D pattern at 0.87 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 0.87 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 3.3 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 2 Cont.
<p><span class="html-italic">Tetrapterys macrocarpa</span> Malpighiaceae (elliptic) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> parameter, (<b>c</b>) 3D pattern at 0.87 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 0.87 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 3.3 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 3
<p><span class="html-italic">Sarcorhachis naranjoana</span> Piperaceae (ovate) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> parameter, (<b>c</b>) 3D pattern at 1.05 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 1.05 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 3.09 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 4
<p><span class="html-italic">Cissampelos owariensis</span> Menispermaceae (circular base) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D pattern at 1.2 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 1.2 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 6.27 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 4 Cont.
<p><span class="html-italic">Cissampelos owariensis</span> Menispermaceae (circular base) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D pattern at 1.2 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 1.2 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 6.27 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 5
<p><span class="html-italic">Paranomus spectrum</span> Proteaceae (obovate) (<b>a</b>) Front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D radiation pattern at 2.34 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 2.34 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 5.31 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 6
<p><span class="html-italic">Cesearia ilicifolia</span> Salicaceae (toothed) (<b>a</b>) Front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D radiation pattern at 5.91 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 5.91 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 6.51 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 7
<p><span class="html-italic">Liquidambar styraciflua</span> Hamamelidaceae (lobed leaves) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D pattern at 4.23 GHz, and cuts of the pattern at <span class="html-italic">f</span> = 4.23 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 5.58 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 7 Cont.
<p><span class="html-italic">Liquidambar styraciflua</span> Hamamelidaceae (lobed leaves) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D pattern at 4.23 GHz, and cuts of the pattern at <span class="html-italic">f</span> = 4.23 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 5.58 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 8
<p><span class="html-italic">Quercus alba</span> Fagaceae (pinnately lobed) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D radiation pattern at 3.18 GHz, and cuts of the pattern at <span class="html-italic">f</span> = 3.18 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 5.88 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">
19 pages, 24057 KiB  
Article
A Four Element Stringray-Shaped MIMO Antenna System for UWB Applications
by Hüseyin Şerif Savcı
Micromachines 2023, 14(10), 1944; https://doi.org/10.3390/mi14101944 - 18 Oct 2023
Cited by 6 | Viewed by 1503
Abstract
This paper presents a CoPlanar-Waveguide (CPW)-fed stingray-shaped Ultra-WideBand (UWB) Multiple-Input–Multiple-Output (MIMO) antenna system designed for microwave imaging applications. Featuring a diagonal square with four inner lines and a vertical line at the center from toe to tip with a CPW feed line, the [...] Read more.
This paper presents a CoPlanar-Waveguide (CPW)-fed stingray-shaped Ultra-WideBand (UWB) Multiple-Input–Multiple-Output (MIMO) antenna system designed for microwave imaging applications. Featuring a diagonal square with four inner lines and a vertical line at the center from toe to tip with a CPW feed line, the unit antenna element looks like a stingray fish skeleton and is, therefore, named as a stingray-shaped antenna. It offers a bandwidth spanning from 3.8 to 12.7 GHz. Fabricated on a 31mil RO5880 RF teflon substrate with a relative permittivity of 2.2, the proposed antenna has dimensions of 26 × 29 × 0.787 mm3. The maximum realized gain achieved is 3.5 dBi with stable omnidirectional radiation patterns. The antenna element is used in a four-antenna MIMO configuration with an isolation-improving structure at the center. The MIMO system has dimensions of 58 × 58 × 0.787 mm3 with a maximum realized gain of 5.3 dBi. The antenna’s performance in terms of MIMO parameters like Envelope Correlation Coefficient (ECC) and Diversity Gain (DG) is within satisfactory limits for medical imaging applications. Time domain analysis also yields positive results, allowing its integration into a breast phantom tumor detection simulation. The simulation and measurement results demonstrate excellent agreement, making this antenna a promising candidate for microwave imaging and biomedical applications. Full article
(This article belongs to the Special Issue Advances in Microwave/Millimeter-Wave Devices and Antennas)
Show Figures

Figure 1

Figure 1
<p>Near-field microwave imaging setup with a tumor inside beast phantom (<b>a</b>) with twelve single antenna elements, (<b>b</b>) with four MIMO antenna systems.</p>
Full article ">Figure 2
<p>Proposed CPW-fed UWB Antenna.</p>
Full article ">Figure 3
<p>Design evolution (<b>a</b>) stage 1, (<b>b</b>) stage 2, (<b>c</b>) proposed, (<b>d</b>) reflection coefficient.</p>
Full article ">Figure 4
<p>Single antenna element (<b>a</b>) parametric analysis of ground arc length, <math display="inline"><semantics> <msub> <mi>R</mi> <mi>r</mi> </msub> </semantics></math> (<b>b</b>) parametric analysis of feedline length, <math display="inline"><semantics> <msub> <mi>L</mi> <mi>f</mi> </msub> </semantics></math> (<b>c</b>) radiation efficiency, total efficiency and gain over frequency (<b>d</b>) reflection coefficient in dB (<b>e</b>) prototype photo (<b>f</b>) s-parameter measurement setup.</p>
Full article ">Figure 5
<p>Time-domain response of single antenna element (<b>a</b>) face-to-face configuration (<b>b</b>) side-by-side configuration (<b>c</b>) face-to-face response (<b>d</b>) side-by-side response.</p>
Full article ">Figure 6
<p>Group delay of UWB antenna.</p>
Full article ">Figure 7
<p>Four element MIMO antennas (<b>a</b>) simple configuration (<b>b</b>) with isolating structure at the center.</p>
Full article ">Figure 8
<p>The scattering parameters of four element MIMO antennas (<b>a</b>,<b>b</b>) simple configuration (<b>c</b>,<b>d</b>) with isolating structure at the center.</p>
Full article ">Figure 8 Cont.
<p>The scattering parameters of four element MIMO antennas (<b>a</b>,<b>b</b>) simple configuration (<b>c</b>,<b>d</b>) with isolating structure at the center.</p>
Full article ">Figure 9
<p>The surface current density for the four elements MIMO antennas (<b>a</b>) at 4.5 GHz without isolating element (<b>b</b>) at 4.5 GHz with isolating element (<b>c</b>) at 6.7 GHz without isolating element (<b>d</b>) at 6.7 GHz with isolating element (<b>e</b>) at 15.5 GHz without isolating element (<b>f</b>) at 15.5 GHz with isolating element.</p>
Full article ">Figure 10
<p>Four-element MIMO antenna system with isolating structure at the center.</p>
Full article ">Figure 11
<p>(<b>a</b>) Measured reflection coefficient, (<b>b</b>) Measured port-to-port isolation.</p>
Full article ">Figure 12
<p>2D Radiation Patterns (<b>a</b>) 6.7 GHz Ant 1, (<b>b</b>) 6.7 GHz Ant 2, (<b>c</b>) 6.7 GHz Ant 3, (<b>d</b>) 6.7 GHz Ant 4, (<b>e</b>) 11 GHz Ant 1, (<b>f</b>) 11 GHz Ant 2, (<b>g</b>) 11 GHz Ant 3, (<b>h</b>) 11 GHz Ant 4.</p>
Full article ">Figure 12 Cont.
<p>2D Radiation Patterns (<b>a</b>) 6.7 GHz Ant 1, (<b>b</b>) 6.7 GHz Ant 2, (<b>c</b>) 6.7 GHz Ant 3, (<b>d</b>) 6.7 GHz Ant 4, (<b>e</b>) 11 GHz Ant 1, (<b>f</b>) 11 GHz Ant 2, (<b>g</b>) 11 GHz Ant 3, (<b>h</b>) 11 GHz Ant 4.</p>
Full article ">Figure 13
<p>MIMO parameters: (<b>a</b>) ECC simulation (with ISO), (<b>b</b>) ECC measurement (with ISO), (<b>c</b>) DG (with ISO), (<b>d</b>) ECC (without ISO), (<b>e</b>) CC (with ISO).</p>
Full article ">Figure 13 Cont.
<p>MIMO parameters: (<b>a</b>) ECC simulation (with ISO), (<b>b</b>) ECC measurement (with ISO), (<b>c</b>) DG (with ISO), (<b>d</b>) ECC (without ISO), (<b>e</b>) CC (with ISO).</p>
Full article ">
25 pages, 21439 KiB  
Article
Accuracy vs. Energy: An Assessment of Bee Object Inference in Videos from On-Hive Video Loggers with YOLOv3, YOLOv4-Tiny, and YOLOv7-Tiny
by Vladimir A. Kulyukin and Aleksey V. Kulyukin
Sensors 2023, 23(15), 6791; https://doi.org/10.3390/s23156791 - 29 Jul 2023
Cited by 10 | Viewed by 2005
Abstract
A continuing trend in precision apiculture is to use computer vision methods to quantify characteristics of bee traffic in managed colonies at the hive’s entrance. Since traffic at the hive’s entrance is a contributing factor to the hive’s productivity and health, we assessed [...] Read more.
A continuing trend in precision apiculture is to use computer vision methods to quantify characteristics of bee traffic in managed colonies at the hive’s entrance. Since traffic at the hive’s entrance is a contributing factor to the hive’s productivity and health, we assessed the potential of three open-source convolutional network models, YOLOv3, YOLOv4-tiny, and YOLOv7-tiny, to quantify omnidirectional traffic in videos from on-hive video loggers on regular, unmodified one- and two-super Langstroth hives and compared their accuracies, energy efficacies, and operational energy footprints. We trained and tested the models with a 70/30 split on a dataset of 23,173 flying bees manually labeled in 5819 images from 10 randomly selected videos and manually evaluated the trained models on 3600 images from 120 randomly selected videos from different apiaries, years, and queen races. We designed a new energy efficacy metric as a ratio of performance units per energy unit required to make a model operational in a continuous hive monitoring data pipeline. In terms of accuracy, YOLOv3 was first, YOLOv7-tiny—second, and YOLOv4-tiny—third. All models underestimated the true amount of traffic due to false negatives. YOLOv3 was the only model with no false positives, but had the lowest energy efficacy and highest operational energy footprint in a deployed hive monitoring data pipeline. YOLOv7-tiny had the highest energy efficacy and the lowest operational energy footprint in the same pipeline. Consequently, YOLOv7-tiny is a model worth considering for training on larger bee datasets if a primary objective is the discovery of non-invasive computer vision models of traffic quantification with higher energy efficacies and lower operational energy footprints. Full article
(This article belongs to the Special Issue Sensor and AI Technologies in Intelligent Agriculture)
Show Figures

Figure 1

Figure 1
<p><b>Left</b>: An on-hive BeePi logger on top of a 2-super Langstroth hive in Logan, Utah; bottom to top: (1) a landing pad; (2) a light gray super; (3) a blue super; (4) a white super with the BeePi logger hardware (see the left image in <a href="#sensors-23-06791-f002" class="html-fig">Figure 2</a>); (5) a waterproof plastic box with a Pi camera inside (see <a href="#sensors-23-06791-f003" class="html-fig">Figure 3</a>) looking down on the landing pad; the box is attached to the front of the third super with two screwed metallic brackets; (6) a wooden migratory hive lid on top of the third white super. <b>Right</b>: two BeePi loggers on top of two super Langstroth hives in Tucson, Arizona; the top boxes on hives contain the logger hardware; water- and dustproof boxes on top of the second supers protect the cameras against rain and dust storms frequent in that area of Arizona.</p>
Full article ">Figure 2
<p><b>Left</b>: BeePi logger hardware; bottom to top: a Raspberry Pi computer coupled to an 8-megapixel Pi camera (see left image in <a href="#sensors-23-06791-f003" class="html-fig">Figure 3</a>); a five terabyte USB disk for archiving data for redundancy in case of GPU computer failures or power supply disruptions; a Pi power charger plugged into a waterproof power cord; videos are wirelessly transferred to a GPU computer over an ad hoc 802.11 local network, where they are processed and archived for redundancy in case of logger storage failures. <b>Right</b>: GEFORCE GTX-980 GPU computer (Arc: x86_64; CPU Family: 6; Model: 60; Model Name: Intel(R) Core(TM) i7-4790K; CPU at 4.00 GHz; BogoMips: 7999.890) with Ubuntu 18.04; all YOLO models were trained on this computer and evaluated for their energy efficacy and operational energy footprint.</p>
Full article ">Figure 3
<p><b>Left</b>: A low-end, low-energy 8-megapixel Raspberry Pi camera v2.1 inside a waterproof camera protection box attached to the front of the super with the on-hive BeePi logger shown in the right picture. <b>Right</b>: An on-hive BeePi logger on top of a one super hive in Logan, Utah in May 2023; bottom to top: (1) a bottom board with a landing pad; (2) a light gray super; (3) a white super with the logger hardware shown in the left image of <a href="#sensors-23-06791-f002" class="html-fig">Figure 2</a>, and a white waterproof camera protection box with the Pi camera in the left image; (4) a white telescoping hive lid.</p>
Full article ">Figure 4
<p>Manually labeled flying bees in a frame from a video captured by a BeePi logger on top of a 2-super hive at an apiary in Tucson, Arizona in September 2022. Each bee object is marked with a rectangle with corners accentuated by small filled light-green circles. When we were not sure if an object is a bee shadow or a bee, we left it unlabeled. Nor did we label any partial bee objects (1/2 bee, 1/3 bee, etc.) or any bee object that we could not recognize, e.g., due to the fogging of a camera lens or video flicker caused by wind. Since our objective was to develop YOLO models to quantify omnidirectional traffic, we avoided, to the best of our visual ability and judgment, labeling stationary bees. In particular, in the above image, all stationary bees on the landing pad are left unlabeled because they do not contribute to traffic.</p>
Full article ">Figure 5
<p><b>Top</b>: Category BTP_BFLMTP (Bee True Positive and Bee Flight Motion True Positive). <b>Bottom</b>: Categories BTP_BCRMTP (Bee True Positive and Bee Crawling Motion True Positive) and BTP_BMTN (Bee True Positive and Bee Motion True Negative).</p>
Full article ">Figure 6
<p><b>Top</b>: categories BTP_BMFN (Bee True Positive and Bee Motion False Negative). <b>Bottom</b>: categories BFN_BFLMTP (Bee False Negative and Bee Flight Motion True Positive) and BTP_BCRMTP (Bee True Positive and Bee Crawling Motion True Positive).</p>
Full article ">Figure 7
<p><b>Top</b>: category BFN_BFLMFN (Bee False Negative and Bee Flight Motion False Negative). <b>Bottom</b>: category BFP (Bee False Positive).</p>
Full article ">Figure 8
<p><b>Left</b>: A Gardner Bender(TM) PM3000 power meter plugged into a wall outlet. <b>Right</b>: A GEFORCE GTX-980 GPU computer plugged into a PM3000 power meter.</p>
Full article ">
14 pages, 6417 KiB  
Article
Development of IoT-Based Real-Time Fire Detection System Using Raspberry Pi and Fisheye Camera
by Chung-Hyun Lee, Woo-Hyuk Lee and Sung-Min Kim
Appl. Sci. 2023, 13(15), 8568; https://doi.org/10.3390/app13158568 - 25 Jul 2023
Cited by 7 | Viewed by 4521
Abstract
In this study, an IoT-based fire detection system was developed to detect and prevent damage from forest fires at an early stage. In Korea, forest fires spread quickly due to the dry climate and winds in spring and autumn, so quick detection and [...] Read more.
In this study, an IoT-based fire detection system was developed to detect and prevent damage from forest fires at an early stage. In Korea, forest fires spread quickly due to the dry climate and winds in spring and autumn, so quick detection and prevention is necessary. To quickly detect and prevent forest fires that occur periodically, a real-time fire detection system was developed by combining a Raspberry Pi and a fisheye camera. A lens with a 220° angle of view was installed, and an image analysis algorithm was implemented using the OpenCV library. The location of the fire was estimated by calculating the polar coordinates of the omnidirectional images. Using the Wi-Fi communication function of the Raspberry Pi, the acquired continuous images were transmitted to the Firebase database, and the images were analyzed to identify the movement path of the forest fire. The developed system was applied to a mountainous area near the Samcheok Campus of Kangwon National University. As a result of the experiment, when the location of points about 25.9 m (average) away from the observation point was predicted, the positional error was analyzed to be about 1.1 m. If the system is improved in the future, it is expected that it will be able to contribute to the early prevention of forest fires with fast and accurate responses. Full article
(This article belongs to the Section Earth Sciences)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Orthoimage, (<b>b</b>) DSM, and (<b>c</b>) 3-D image of Bukjeong Mountain in Samcheok, Gangwon-do.</p>
Full article ">Figure 2
<p>(<b>a</b>) Concept of real-time forest fire detection system combining Raspberry Pi and fisheye camera, and (<b>b</b>) photo of the system mounted on a tripod.</p>
Full article ">Figure 3
<p>(<b>a</b>) Polar image taken using a fisheye camera for a (<b>b</b>) transparent hemisphere with latitude indicated.</p>
Full article ">Figure 4
<p>(<b>a</b>) Fire detection in a picture containing fire and red objects (day), and (<b>b</b>) fire detection in a picture containing fire and bright lights (night).</p>
Full article ">Figure 5
<p>An example of (<b>a</b>) a topographical map where two forest fires occurred, (<b>b</b>) the position of the forest fire in the polar image obtained through the fisheye camera, and (<b>c</b>) a cross-sectional diagram from the observation point to the forest fire point.</p>
Full article ">Figure 6
<p>Estimation of a suitable location to install the developed system. (<b>a</b>) The visibility index result, where the red circle is the point where the value is high and the triangle is the point where the value is low, (<b>b</b>) the viewshed analysis result at the point where the visibility index is high, and (<b>c</b>) the result at the point where the visibility index is low.</p>
Full article ">Figure 7
<p>(<b>a</b>) A picture of the developed system installed in the study area, (<b>b</b>) a polar image taken through the system, and images taken with a normal camera facing (<b>c</b>) forward and (<b>d</b>) vertically upward.</p>
Full article ">Figure 8
<p>Some of the images continuously acquired during (<b>a</b>) the day and (<b>b</b>) night, respectively, through the developed system.</p>
Full article ">Figure 8 Cont.
<p>Some of the images continuously acquired during (<b>a</b>) the day and (<b>b</b>) night, respectively, through the developed system.</p>
Full article ">Figure 9
<p>GPS-recorded positions of the subject (day: yellow, night: green) and calculated positions from the polar image (white dots). (<b>a</b>) daytime experiment, (<b>b</b>) nighttime experiment.</p>
Full article ">Figure 10
<p>Image of the point where the GPS of the subject was recorded on DSM(Digital Elevation Model). (<b>a</b>) daytime experiment, (<b>b</b>) nighttime experiment.</p>
Full article ">
12 pages, 2820 KiB  
Perspective
Design and Fabrication of Broadband InGaAs Detectors Integrated with Nanostructures
by Bo Yang, Yizhen Yu, Guixue Zhang, Xiumei Shao and Xue Li
Sensors 2023, 23(14), 6556; https://doi.org/10.3390/s23146556 - 20 Jul 2023
Cited by 6 | Viewed by 2696
Abstract
A visible–extended shortwave infrared indium gallium arsenide (InGaAs) focal plane array (FPA) detector is the ideal choice for reducing the size, weight and power (SWaP) of infrared imaging systems, especially in low-light night vision and other fields that require simultaneous visible and near-infrared [...] Read more.
A visible–extended shortwave infrared indium gallium arsenide (InGaAs) focal plane array (FPA) detector is the ideal choice for reducing the size, weight and power (SWaP) of infrared imaging systems, especially in low-light night vision and other fields that require simultaneous visible and near-infrared light detection. However, the lower quantum efficiency in the visible band has limited the extensive application of the visible–extended InGaAs FPA. Recently, a novel optical metasurface has been considered a solution for a high-performance semiconductor photoelectric device due to its highly controllable property of electromagnetic wave manipulation. Broadband Mie resonator arrays, such as nanocones and nanopillars designed with FDTD methods, were integrated on a back-illuminated InGaAs FPA as an AR metasurface. The visible–extended InGaAs detector was fabricated using substrate removal technology. The nanostructures integrated into the Vis-SWIR InGaAs detectors could realize a 10–20% enhanced quantum efficiency and an 18.8% higher FPA response throughout the wavelength range of 500–1700 nm. Compared with the traditional AR coating, nanostructure integration has advantages, such as broadband high responsivity and omnidirection antireflection, as a promising route for future Vis-SWIR InGaAs detectors with higher image quality. Full article
(This article belongs to the Special Issue Semiconductor Sensors towards Optoelectronic Device Applications)
Show Figures

Figure 1

Figure 1
<p>The fabrication flow of the broadband InGaAs detector.</p>
Full article ">Figure 2
<p>Schematic of textured InGaAs focal plane arrays: (<b>a</b>) SiNx nanocone, (<b>b</b>) InP nanopillar, (<b>c</b>) InP nanocone.</p>
Full article ">Figure 3
<p>A flowchart of the fabrication process for InP nanostructure arrays with three different masks: (<b>a</b>) photoresist, (<b>b</b>) metal, (<b>c</b>) silicon nitride.</p>
Full article ">Figure 4
<p>Responsivity of the detectors with and without the AR film.</p>
Full article ">Figure 5
<p>The reflectivity simulation of the nanocone with different heights.</p>
Full article ">Figure 6
<p>Reflectivity of the InGaAs FPAs with different antireflection systems.</p>
Full article ">Figure 7
<p>The changes in the reflection of nanocone structure at varied incident angles (<b>left</b>) and the comparison of the reflectivity of the different structures vs. angle of incidence at the wavelength of 1550 nm (<b>right</b>).</p>
Full article ">Figure 8
<p>The morphology of InP nanopillars prepared by etching with mask of photoresist (<b>a</b>), Cr (<b>b</b>) and SiN<sub>x</sub> (<b>c</b>).</p>
Full article ">Figure 9
<p>Simulated quantum efficiency of metasurface-integrated (<b>a</b>) SWIR InGaAs FPA, (<b>b</b>) Vis-SWIR InGaAs FPA, (<b>c</b>) ultrathin InP Vis-SWIR InGaAs FPA and (<b>d</b>) measured quantum efficiency of the InGaAs FPAs with different systems.</p>
Full article ">
Back to TopTop