[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (114)

Search Parameters:
Keywords = omnidirectional image

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 5968 KiB  
Article
The Optimization of PID Controller and Color Filter Parameters with a Genetic Algorithm for Pineapple Tracking Using an ROS2 and MicroROS-Based Robotic Head
by Carolina Maldonado-Mendez, Sergio Fabian Ruiz-Paz, Isaac Machorro-Cano, Antonio Marin-Hernandez and Sergio Hernandez-Mendez
Computation 2025, 13(3), 69; https://doi.org/10.3390/computation13030069 - 7 Mar 2025
Viewed by 42
Abstract
This work proposes a vision system mounted on the head of an omnidirectional robot to track pineapples and maintain them at the center of its field of view. The robot head is equipped with a pan–tilt unit that facilitates dynamic adjustments. The system [...] Read more.
This work proposes a vision system mounted on the head of an omnidirectional robot to track pineapples and maintain them at the center of its field of view. The robot head is equipped with a pan–tilt unit that facilitates dynamic adjustments. The system architecture, implemented in Robot Operating System 2 (ROS2), performs the following tasks: it captures images from a webcam embedded in the robot head, segments the object of interest based on color, and computes its centroid. If the centroid deviates from the center of the image plane, a proportional–integral–derivative (PID) controller adjusts the pan–tilt unit to reposition the object at the center, enabling continuous tracking. A multivariate Gaussian function is employed to segment objects with complex color patterns, such as the body of a pineapple. The parameters of both the PID controller and the multivariate Gaussian filter are optimized using a genetic algorithm. The PID controller receives as input the (x, y) positions of the pan–tilt unit, obtained via an embedded board and MicroROS, and generates control signals for the servomotors that drive the pan–tilt mechanism. The experimental results demonstrate that the robot successfully tracks a moving pineapple. Additionally, the color segmentation filter can be further optimized to detect other textured fruits, such as soursop and melon. This research contributes to the advancement of smart agriculture, particularly for fruit crops with rough textures and complex color patterns. Full article
Show Figures

Figure 1

Figure 1
<p>Omnidirectional robot used in the experiments. The robot’s head, located at the top, where a webcam is mounted on the pan–tilt unit.</p>
Full article ">Figure 2
<p>The two images on the left show the robot head mounted on an PTU, with a webcam housed inside. The angular positions of the PTU are obtained using an Arduino Due board. The image on the right illustrates PTU, which consists of two motors: one for movement along the <span class="html-italic">x</span>-axis and another for movement along the <span class="html-italic">y</span>-axis.</p>
Full article ">Figure 3
<p>The architecture implemented in ROS2 consists of both hardware and software components. (1) Hardware: This includes input and output devices such as a servo motor, a webcam, and an Arduino Due board with microROS to control the servo motors responsible for moving the pan–tilt unit (PTU). (2) Software: The system is structured into modules with specific functionalities, including object segmentation, centroid calculation, and tracking using two PID controllers. The servomotors are powered by lithium batteries. ROS2 manages the communication between the software and hardware layers.</p>
Full article ">Figure 4
<p>Communication graph of nodes and topics in ROS2.</p>
Full article ">Figure 5
<p>A total of 35 photos of pineapples were used to train the MGF and optimize its parameters with the GA.</p>
Full article ">Figure 6
<p>In the upper part, 3 images are shown where the body of the pineapple are manually colored in black. The lower part shows the images generated in MATLAB, where the black pixels are changed to white, and the remaining pixels are colored black.</p>
Full article ">Figure 7
<p>The segmented pineapples are shown using the best individual from the GA.</p>
Full article ">Figure 8
<p>The image processing steps applied to calculate the centroid of the pineapple body are shown: (<b>a</b>) the original image, (<b>b</b>) the segmented image with the MGF filter optimized with the GA, (<b>c</b>) the image resulting from the application of the smoothing filter, (<b>d</b>) the image after contour detection, and (<b>e</b>) the blue point indicating the centroid of the pineapple contour.</p>
Full article ">Figure 9
<p>The upper images depict pineapples captured in an outdoor environment, while the lower images display the segmented pineapple highlighted with a green outline.</p>
Full article ">Figure 10
<p>Images of pineapples with green hues captured in an outdoor environment [<a href="#B35-computation-13-00069" class="html-bibr">35</a>]. The green outline indicates the segmentation of a pineapple, which is an incorrect detection.</p>
Full article ">Figure 11
<p>Images of pineapples with shades similar to those used during training taken from [<a href="#B35-computation-13-00069" class="html-bibr">35</a>]. The green circle highlights the segmented pineapple, while the blue dot represents the centroid of the segmented area. In (<b>b</b>,<b>c</b>,<b>e</b>,<b>g</b>,<b>h</b>), the segmentation results are accurate. In (<b>a</b>), only the pineapple body with yellow hues was segmented. In (<b>d</b>,<b>f</b>), partial segmentation of the pineapple body is observed due to occlusions in these images.</p>
Full article ">Figure 12
<p>The visual information is processed to segment the object of interest, and this centroid is calculated. Two PID controllers in the pan and tilt unit are used, in order to track the object with a minimum error and soft movement. For this purpose, the centroid of the object is located at the center of the plane image.</p>
Full article ">Figure 13
<p>The images show the green square placed at the bottom of the board. In (<b>a</b>,<b>c</b>), the centroid of the green object does not coincide with the center of the image plane; in (<b>b</b>,<b>d</b>), the control action was applied, and the centroid of the green object was positioned near the center of the image plane. The purple circle indicates the segmented object.</p>
Full article ">Figure 14
<p>It shows the change in the average error for both pan and tilt across the 20 repetitions.</p>
Full article ">Figure 15
<p>Tracking the green object with the robot head. The purple circle indicates the segmented object.</p>
Full article ">Figure 16
<p>It shows the change in the average error in the pan and tilt of the 20 repetitions by following the moving green object. The PID was optimized with the GA.</p>
Full article ">Figure 17
<p>It shows the change in the average error in the pan and tilt of the 20 repetitions by following the moving green object.</p>
Full article ">Figure 18
<p>The images show the pineapple placed on a table. In the images of the left column, the robot head does not observe the centroid of the pineapple in the center of its image plane. In the images in the right column, the control action is observed to have been performed, and the centroid of the pineapple is close to the center of the image plane. The purple circle indicates the segmented object.</p>
Full article ">Figure 19
<p>It shows the change in the average error in the pan and tilt of the 20 repetitions. The PID was optimized manually.</p>
Full article ">Figure 20
<p>Tracking the pineapple with the robot head. The purple circle indicates the segmented object.</p>
Full article ">Figure 21
<p>The figure shows the change in the average error in the pan and tilt across the 20 repetitions while tracking the moving pineapple.</p>
Full article ">
22 pages, 11585 KiB  
Article
Marine Radar Target Ship Echo Generation Algorithm and Simulation Based on Radar Cross-Section
by Chang Li, Xiao Yang, Hongxiang Ren, Shihao Li and Xiaoyu Feng
J. Mar. Sci. Eng. 2025, 13(2), 348; https://doi.org/10.3390/jmse13020348 - 14 Feb 2025
Viewed by 300
Abstract
In this study, a simplified radar echo signal model suitable for radar simulators and a Radar Cross-Section (RCS) calculation model based on the Physical Optics (PO) method was developed. A comprehensive radar target ship echo generation algorithm was designed, and the omnidirectional radar [...] Read more.
In this study, a simplified radar echo signal model suitable for radar simulators and a Radar Cross-Section (RCS) calculation model based on the Physical Optics (PO) method was developed. A comprehensive radar target ship echo generation algorithm was designed, and the omnidirectional radar RCS values of three typical ships were calculated. The simulation generates radar target ship echo images under varying incident angles (0–360°), detection distances (0–24 nautical miles), and three common target material properties. The simulation results, compared with those from existing radar simulators and real radar systems, show that the method proposed in this study, based on RCS values, generates highly realistic radar target ship echoes. It accurately simulates radar echoes under different target ship headings, distances, and material influences, fully meeting the technical requirements of the STCW international convention for radar simulators. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Figure 1
<p>Target ship echo generation algorithm.</p>
Full article ">Figure 2
<p>Vincent formula diagram.</p>
Full article ">Figure 3
<p>Incident angle calculation algorithm.</p>
Full article ">Figure 4
<p>Electromagnetic calculation and analysis process of the target ship.</p>
Full article ">Figure 5
<p>Three-dimensional geometric model of the ship.</p>
Full article ">Figure 6
<p>Omnidirectional RCS diagram of the ship.</p>
Full article ">Figure 7
<p>RCS line graph of the bulk carrier.</p>
Full article ">Figure 8
<p>RCS line graph of the container ship.</p>
Full article ">Figure 9
<p>RCS line graph of the fishing boat.</p>
Full article ">Figure 10
<p>Echo diagrams of the bulk carrier at different angles.</p>
Full article ">Figure 11
<p>Echo diagrams of the container ship at different angles.</p>
Full article ">Figure 12
<p>Echo images of the target ship at different distances.</p>
Full article ">Figure 13
<p>RCS Comparison of Ships with Three Different Materials.</p>
Full article ">Figure 14
<p>Target ship echo images of different materials.</p>
Full article ">Figure 15
<p>Echo images from the different simulators.</p>
Full article ">Figure 16
<p>Enlarged echo image of the target ship.</p>
Full article ">Figure 17
<p>Comparison between the simulator and real machine.</p>
Full article ">
45 pages, 10284 KiB  
Article
Guided Wave Propagation in a Realistic CFRP Fuselage Panel: Proof of Concept for Early-Stage Damage Detection
by Fatma Sellami, Vittorio Memmolo and Mirko Hornung
Sensors 2025, 25(4), 1104; https://doi.org/10.3390/s25041104 - 12 Feb 2025
Viewed by 520
Abstract
This paper presents an experimental study of wave motions and a global diagnostics method in a realistic aerospace-grade composite component with a complex design. Due to the frequency dependence of the velocity, wave propagation in anisotropic materials is difficult to describe quantitatively. The [...] Read more.
This paper presents an experimental study of wave motions and a global diagnostics method in a realistic aerospace-grade composite component with a complex design. Due to the frequency dependence of the velocity, wave propagation in anisotropic materials is difficult to describe quantitatively. The analysis of experimental ultrasonic wave propagation and the interactions with discontinuities in thin-walled aircraft structures can provide a plethora of information on the wave structure, the mode shapes, and stiffness reduction. An experiment is devised with a network of various omnidirectional sensor configurations to activate and measure structural responses. The measurement process can be leveraged for flaw detection in large multilayered structures. Physically, this corresponds to analyzing the dispersive behavior and scattering properties of ultrasonic waves, the shape of the waveforms, and their corresponding velocities. Ultrasonic waves are measured in a realistic CFRP fuselage panel in a pristine state and after impacts at different energy levels. Simulations do not allow the wave motion in complex and large design structures to be entirely comprehended. The sensitivity of the guided waves as a damage detection tool is proved for the fuselage structure by an extensive measurement campaign and a probabilistic imaging approach based on health indicator fusion. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

Figure 1
<p>Schematic of residual strain versus energy level and the derived design constraints [<a href="#B63-sensors-25-01104" class="html-bibr">63</a>].</p>
Full article ">Figure 2
<p>Schematic of scattering factors constraining the composite design (<b>a</b>) and typical design stress–strain allowables for CFRP (<b>b</b>) [<a href="#B63-sensors-25-01104" class="html-bibr">63</a>].</p>
Full article ">Figure 3
<p>Sensor networks adopted for SHM. The numbers (in black for Network No. 1 and in blue for Network No. 2) identify the sensors.</p>
Full article ">Figure 4
<p>Network No. 3: Network No. 2 with a higher sensor coverage. The numbers identify the sensors.</p>
Full article ">Figure 5
<p>Example of a sensor network.</p>
Full article ">Figure 6
<p>The investigated CFRP fuselage panel with the five sections.</p>
Full article ">Figure 7
<p>Data acquisition system.</p>
Full article ">Figure 8
<p>Experimental set-up.</p>
Full article ">Figure 9
<p>Piezoelectric transducer array and resulting propagation paths where damage indices are calculated (<b>a</b>). Structural mesh where damage probability index is derived using probability distribution function (<b>b</b>) [<a href="#B74-sensors-25-01104" class="html-bibr">74</a>].</p>
Full article ">Figure 10
<p>Dispersion curves for the propagation direction 0° displaying first modes: S0 (symmetric), S0’ (shear) and A0 (anti-symmetric).</p>
Full article ">Figure 11
<p>Group velocities in all four propagation directions.</p>
Full article ">Figure 12
<p>A typical 5-count burst actuation waveform at 190 kHz (<b>left</b>) and a receiver signal (<b>right</b>).</p>
Full article ">Figure 13
<p>Experimental tuning process: (<b>top</b>) peak voltages over a frequency range, (<b>bottom</b>) group velocities in three sections of the panel.</p>
Full article ">Figure 14
<p>Group velocity: (<b>top</b>) Experimental versus analytical wave velocities for wave Path 2 to 14, (<b>bottom</b>) group velocity measurements in Sections 3 and 5.</p>
Full article ">Figure 15
<p>Received signals as a function of frequency.</p>
Full article ">Figure 16
<p>Waveforms in network no. 3: In Sections 3 and 5, crossing Sections 3 to 4, 4 to 5, and 3 to 5.</p>
Full article ">Figure 17
<p>Waveforms in Section 1.</p>
Full article ">Figure 18
<p>Wave Paths 2 to 11 and 2 to 25.</p>
Full article ">Figure 19
<p>Waveforms in Sections 3 and 5 at a center frequency of 250 kHz: (<b>top</b>) experimental received signal, (<b>bottom</b>) first direct wave.</p>
Full article ">Figure 20
<p>Effects on waves’ phase and amplitude.</p>
Full article ">Figure 21
<p>Waveforms after a series of impacts: The first wave packet.</p>
Full article ">Figure 22
<p>The polar pattern of experimental group velocity at 250 kHz. The nominal scale (<b>a</b>) and enlarged scale (<b>b</b>) used to highlight the variability over the propagation direction.</p>
Full article ">Figure 23
<p>Damage probability index maps relying on an energy-based index. The damage map (<b>a</b>) and contour plot (<b>b</b>) are obtained using the linear weight distribution function. Colorbar in the range [0–1]. SHM carried out at 250 kHz.</p>
Full article ">Figure 24
<p>Damage probability index maps relying on energy (<b>a</b>), RMSD (<b>b</b>), and correlation coefficient (<b>c</b>) indices. Colorbar in the range [0–1]. SHM carried out at 250 kHz.</p>
Full article ">Figure 25
<p>Damage probability index maps relying on a multi-parameter approach. Damage map and contour plot are obtained using linear (<b>a</b>), modified linear (<b>b</b>), and elliptical (<b>c</b>) weight distribution functions. Colorbar in the range [0–1]. SHM carried out at 250 kHz.</p>
Full article ">Figure 26
<p>Damage probability index maps relying on an energy-based index. Damage map and contour plot are obtained by applying 0% (<b>a</b>), 30% (<b>b</b>), and 50% (<b>c</b>) thresholding. Colorbar in the range [0–1]. SHM carried out at 250 kHz.</p>
Full article ">Figure 27
<p>Damage probability index maps relying on an energy-based index. Damage map and contour plot are obtained using 421 paths (<b>a</b>), transducers from Section 1 (<b>b</b>), and 55 paths (<b>c</b>). Colorbar in the range [0–1]. SHM carried out at 250 kHz.</p>
Full article ">Figure 28
<p>Damage probability index maps relying on a few transducers (55 paths). Damage map and contour plot are obtained using energy (<b>a</b>), cross-correlation (<b>b</b>), and RMSD (<b>c</b>) indices. Colorbar in the range [0–1]. SHM carried out at 250 kHz.</p>
Full article ">Figure 29
<p>Damage probability index maps in Section 3 using a populated sensor array. Damage maps are obtained using energy (<b>a</b>), RMSD (<b>b</b>), and multi-parameter (<b>c</b>) indices. Colorbar in the range [0–1]. SHM carried out at 250 kHz.</p>
Full article ">Figure 30
<p>A-scans in Section 1.</p>
Full article ">Figure 31
<p>A-scans in Sections 3 and 5.</p>
Full article ">Figure 32
<p>B-scans in Sections 1, 3, and 5.</p>
Full article ">
23 pages, 21957 KiB  
Article
Terrain Analysis According to Multiscale Surface Roughness in the Taklimakan Desert
by Sebastiano Trevisani and Peter L. Guth
Land 2024, 13(11), 1843; https://doi.org/10.3390/land13111843 - 5 Nov 2024
Viewed by 937
Abstract
Surface roughness, interpreted in the wide sense of surface texture, is a generic term referring to a variety of aspects and scales of spatial variability of surfaces. The analysis of solid earth surface roughness is useful for understanding, characterizing, and monitoring geomorphic factors [...] Read more.
Surface roughness, interpreted in the wide sense of surface texture, is a generic term referring to a variety of aspects and scales of spatial variability of surfaces. The analysis of solid earth surface roughness is useful for understanding, characterizing, and monitoring geomorphic factors at multiple spatiotemporal scales. The different geomorphic features characterizing a landscape exhibit specific characteristics and scales of surface texture. The capability to selectively analyze specific roughness metrics at multiple spatial scales represents a key tool in geomorphometric analysis. This research presents a simplified geostatistical approach for the multiscale analysis of surface roughness, or of image texture in the case of images, that is highly informative and interpretable. The implemented approach is able to describe two main aspects of short-range surface roughness: omnidirectional roughness and roughness anisotropy. Adopting simple upscaling approaches, it is possible to perform a multiscale analysis of roughness. An overview of the information extraction potential of the approach is shown for the analysis of a portion of the Taklimakan desert (China) using a 30 m resolution DEM derived from the Copernicus Glo-30 DSM. The multiscale roughness indexes are used as input features for unsupervised and supervised learning tasks. The approach can be refined both from the perspective of the multiscale analysis as well as in relation to the surface roughness indexes considered. However, even in its present, simplified form, it can find direct applications in relation to multiple contexts and research topics. Full article
(This article belongs to the Section Land, Soil and Water)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Reprojected COP DEM (30 m resolution, UTM F44) of the area of interest overlaid on the hillshade.</p>
Full article ">Figure 2
<p>Sentinel-2 true color RGB image (bands 4, 3, and 2) of the study area, with the main dune morphologies labeled.</p>
Full article ">Figure 3
<p>Main dune morphologies in the study area, visualized using Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>). From top to bottom: network/transverse dunes, longitudinal and transverse dunes, and dome-shaped dunes.</p>
Full article ">Figure 4
<p>Mixed morphologies in the area of interest, visualized using Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>). From top to bottom: outcropping bedrock with shadow and linear dunes, fluvial morphology, and a flat area with minor dune morphologies.</p>
Full article ">Figure 5
<p>RA direction, where the RA strength is higher than 0.3, overlaid on the hillshade (<b>a</b>) and the residual DEM (<b>b</b>) calculated for level L2.</p>
Full article ">Figure 6
<p>Omnidirectional short-range roughness (m) for the different resolutions. Different color scales for each diagram.</p>
Full article ">Figure 7
<p>Roughness anisotropy strength at different resolutions. Different color scales for each diagram.</p>
Full article ">Figure 8
<p>RGB image (each band normalized) of 3 omnidirectional roughness indexes computed at different resolutions (B = L1; G = L2; R = L4). Despite the high correlation of the three indexes, they differentiate very well the morphological features of the area. For example, they markedly highlight the characteristic smoothness of interdune areas of the longitudinal dunes south of the mountain ridge.</p>
Full article ">Figure 9
<p>RGB image (each band normalized) of 3 anisotropy strength roughness indexes computed at different resolutions (B = L1; G = L4; R = L16). In the dune fields north of the mountains, long-wavelength anisotropic features prevail; in contrast, for the southern longitudinal dunes, shorter anisotropic features (L4) are highlighted.</p>
Full article ">Figure 10
<p>Landscape clustered according to multiscale surface roughness indexes. The cluster centers in terms of OR and RA are described in <a href="#land-13-01843-f011" class="html-fig">Figure 11</a>.</p>
Full article ">Figure 11
<p>Cluster centers of the 7 classes resulting from K-means clustering for OR and RA at the different levels.</p>
Full article ">Figure 12
<p>MRI clustering results in the area of the northern dune field, characterized by network and transverse dunes. Clustering results (<b>d</b>), Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>).</p>
Full article ">Figure 13
<p>MRI clustering results in the area of the southern longitudinal dune fields. Clustering results (<b>d</b>), Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>).</p>
Full article ">Figure 14
<p>MRI clustering results in the area with fluvial morphology, outcropping bedrock, and dome dune fields. Clustering results (<b>d</b>), Sentinel-2 imagery (<b>a</b>), hillshade (<b>b</b>), and residual DEM (<b>c</b>).</p>
Full article ">Figure 15
<p>Manual classification of crest lines (<b>a</b>) for large dunes using visual analysis of slope (<b>b</b>), profile curvature (<b>c</b>), and residual DEM (<b>d</b>). Crest lines are associated with high positive profile curvature, strongly positive residual DEM, and low slope. These locations are then located in areas in which the neighborhood is characterized by an abrupt variation in the selected geomorphometric derivatives.</p>
Full article ">Figure 16
<p>Probability of observing a crest obtained by means of RF considering the GDs integrated with the MRIs (<b>a</b>) and only the five GDs (<b>b</b>) to obtain details of the study area, which is located on the western mountain ridge. The RF model integrating the MRIs provides a more focused prediction of crest lines of large dunes. In (<b>c</b>), the prediction of the crest lines of the two RF models is compared. Pixels with a probability higher than 0.8 have been classified as crests. The transparent color is where both models predicted a not-crest pixel, green is where both models predicted a crest, and red and blue are where, respectively, only RF GDs and RF GDs + MRIs predicted a crest.</p>
Full article ">Figure 17
<p>Variables’ importance in the two RF models according to the mean decrease in the Gini index ((<b>a</b>), RF based on GDs; (<b>b</b>), RF based on GDs integrated with MRIs).</p>
Full article ">Figure 18
<p>Prediction of crest lines with the RF model based on GDs and MRIs of an unseen area ((<b>c</b>), green box) external to the one with reference data used for training and testing ((<b>c</b>), red box). The reference crest lines (<b>a</b>) have been manually digitized by means of visual analysis of the profile curvature, the residual DEM, and the slope; the predicted crest lines have been derived as crests of all of the pixels with a probability above 0.8. The predicted crest lines are compared with the reference data (<b>b</b>). Green pixels are correctly classified as crests; red and blue pixels are incorrectly classified, respectively, as crests and not crests.</p>
Full article ">
27 pages, 6983 KiB  
Article
DA-YOLOv7: A Deep Learning-Driven High-Performance Underwater Sonar Image Target Recognition Model
by Zhe Chen, Guohao Xie, Xiaofang Deng, Jie Peng and Hongbing Qiu
J. Mar. Sci. Eng. 2024, 12(9), 1606; https://doi.org/10.3390/jmse12091606 - 10 Sep 2024
Cited by 1 | Viewed by 1725
Abstract
Affected by the complex underwater environment and the limitations of low-resolution sonar image data and small sample sizes, traditional image recognition algorithms have difficulties achieving accurate sonar image recognition. The research builds on YOLOv7 and devises an innovative fast recognition model designed explicitly [...] Read more.
Affected by the complex underwater environment and the limitations of low-resolution sonar image data and small sample sizes, traditional image recognition algorithms have difficulties achieving accurate sonar image recognition. The research builds on YOLOv7 and devises an innovative fast recognition model designed explicitly for sonar images, namely the Dual Attention Mechanism YOLOv7 model (DA-YOLOv7), to tackle such challenges. New modules such as the Omni-Directional Convolution Channel Prior Convolutional Attention Efficient Layer Aggregation Network (OA-ELAN), Spatial Pyramid Pooling Channel Shuffling and Pixel-level Convolution Bilat-eral-branch Transformer (SPPCSPCBiFormer), and Ghost-Shuffle Convolution Enhanced Layer Aggregation Network-High performance (G-ELAN-H) are central to its design, which reduce the computational burden and enhance the accuracy in detecting small targets and capturing local features and crucial information. The study adopts transfer learning to deal with the lack of sonar image samples. By pre-training the large-scale Underwater Acoustic Target Detection Dataset (UATD dataset), DA-YOLOV7 obtains initial weights, fine-tuned on the smaller Smaller Common Sonar Target Detection Dataset (SCTD dataset), thereby reducing the risk of overfitting which is commonly encountered in small datasets. The experimental results on the UATD, the Underwater Optical Target Detection Intelligent Algorithm Competition 2021 Dataset (URPC), and SCTD datasets show that DA-YOLOV7 exhibits outstanding performance, with [email protected] scores reaching 89.4%, 89.9%, and 99.15%, respectively. In addition, the model maintains real-time speed while having superior accuracy and recall rates compared to existing mainstream target recognition models. These findings establish the superiority of DA-YOLOV7 in sonar image analysis tasks. Full article
Show Figures

Figure 1

Figure 1
<p>Structure of the YOLOv7.</p>
Full article ">Figure 2
<p>(<b>left</b>): OA-ELAN structure diagram, (<b>right</b>): ODConv Structure diagram.</p>
Full article ">Figure 3
<p>The CPCA attention mechanism.</p>
Full article ">Figure 4
<p><b>Left</b>: SPPCSPC structure, <b>Right</b>: BiFormer structure.</p>
Full article ">Figure 5
<p>Structure diagram of G-ELAN-H.</p>
Full article ">Figure 6
<p>The DA-YOLOv7 network.</p>
Full article ">Figure 7
<p>Confusion matrix of the ablation model: (<b>a</b>) YOLOv7; (<b>b</b>) YOLOv7 + OA-ELAN; (<b>c</b>) YOLOv7 + OA-ELAN + SPPCSPCBiFormer; (<b>d</b>) DA-YOLOv7.</p>
Full article ">Figure 8
<p>The PR curve: (<b>a</b>) YOLOv7; (<b>b</b>) YOLOv7 + OA-ELAN; (<b>c</b>) YOLOv7 + OA-ELAN + SPPCSPCBiFormer; (<b>d</b>) DA-YOLOv7.</p>
Full article ">Figure 9
<p>Curve of the change in loss value on the UATD dataset.</p>
Full article ">Figure 10
<p>Prediction results of various targets in UATD multi-beam forward-looking sonar images.</p>
Full article ">Figure 11
<p>SCTD sonar image dataset: (<b>a</b>) human; (<b>b</b>) ship; (<b>c</b>) aircraft.</p>
Full article ">Figure 12
<p>Flowchart of the training strategy for the SCTD dataset.</p>
Full article ">Figure 13
<p>The effect of recognition on the SCTD dataset: (<b>a</b>) SCTD mAP Results; (<b>b</b>) SCTD aircraft-AP results; (<b>c</b>) SCTD human-AP results; (<b>d</b>) SCTD ship-AP results.</p>
Full article ">Figure 14
<p>The sample information of URPC is as follows: (<b>a</b>) Labels: The upper left corner shows the distribution of categories; the upper right corner presents the visualization of all box sizes; the lower left corner indicates the distribution of the box centroid position; the lower right corner depicts the distribution of the box aspect ratio. (<b>b</b>) Example images.</p>
Full article ">Figure 15
<p>Recognition results in multiple underwater scenes.</p>
Full article ">Figure 16
<p>RCurve of loss value changes on the UPRC dataset.</p>
Full article ">
25 pages, 28511 KiB  
Article
A Method for Estimating the Distribution of Trachinotus ovatus in Marine Cages Based on Omnidirectional Scanning Sonar
by Yu Hu, Jiazhen Hu, Pengqi Sun, Guohao Zhu, Jialong Sun, Qiyou Tao, Taiping Yuan, Gen Li, Guoliang Pang and Xiaohua Huang
J. Mar. Sci. Eng. 2024, 12(9), 1571; https://doi.org/10.3390/jmse12091571 - 6 Sep 2024
Viewed by 727
Abstract
In order to accurately estimate the distribution of Trachinotus ovatus in marine cages, a novel method was developed using omnidirectional scanning sonar and deep-learning techniques. This method involved differentiating water layers and clustering data layer by layer to achieve precise location estimation. The [...] Read more.
In order to accurately estimate the distribution of Trachinotus ovatus in marine cages, a novel method was developed using omnidirectional scanning sonar and deep-learning techniques. This method involved differentiating water layers and clustering data layer by layer to achieve precise location estimation. The approach comprised two main components: fish identification and fish clustering. Firstly, omnidirectional scanning sonar was employed to perform spiral detection within marine cages, capturing fish image data. These images were then labeled to construct a training dataset for an enhanced CS-YOLOv8s model. After training, the CS-YOLOv8s model was used to identify and locate fish within the images. Secondly, the cages were divided into water layers with depth intervals of 40 cm. The identification coordinate data for each water layer were clustered using the DBSCAN method to generate location coordinates for the fish in each layer. Finally, the coordinate data from all water layers were consolidated to determine the overall distribution of fish within the cage. This method was shown, through multiple experimental results, to effectively estimate the distribution of Trachinotus ovatus in marine cages, closely matching the distributions detected manually. Full article
(This article belongs to the Special Issue New Techniques and Equipment in Large Offshore Aquaculture Platform)
Show Figures

Figure 1

Figure 1
<p>Satellite image map of the experimental site.</p>
Full article ">Figure 2
<p>Aerial view of the experimental site.</p>
Full article ">Figure 3
<p>Cage used in the experiment.</p>
Full article ">Figure 4
<p><span class="html-italic">Trachinotus ovatus</span> used in the experiment. (<b>a</b>) Body length; (<b>b</b>) Body height.</p>
Full article ">Figure 5
<p>Working principle diagram of omnidirectional scanning sonar. (<b>a</b>) Schematic diagram of sonar scanning; (<b>b</b>) Work diagram.</p>
Full article ">Figure 6
<p>Omnidirectional scanning sonar.</p>
Full article ">Figure 7
<p>Sonar Assembly.</p>
Full article ">Figure 8
<p>Side view of the sea cage. The sonar in the yellow box is located on the center axis of the net cage, and the yellow dotted lines shows the range covered by the sonar.</p>
Full article ">Figure 9
<p>Training process (Loss).</p>
Full article ">Figure 10
<p>Training process (mAP@0.5%).</p>
Full article ">Figure 11
<p>Coordinate transformation. (<b>a</b>) A coordinate system with sonar as the origin; (<b>b</b>) A coordinate system with the top left corner as the origin.</p>
Full article ">Figure 12
<p>Water layer division.</p>
Full article ">Figure 13
<p>Clustering in the 100–140 cm water layer. (<b>a</b>) Fish distribution; (<b>b</b>) Cluster effect diagram.</p>
Full article ">Figure 14
<p>Water layer data and the data after clustering. (<b>a</b>) Fish distribution map before clustering; (<b>b</b>) Cluster diagram of noisy points.</p>
Full article ">Figure 15
<p>Center and noise point of core object. (<b>a</b>) Distribution map of center point and noise point; (<b>b</b>) Centralization rendering.</p>
Full article ">Figure 16
<p>Sonar images of different numbers of fish. (<b>a</b>) 100 fish; (<b>b</b>) 150 fish; (<b>c</b>) 200 fish.</p>
Full article ">Figure 17
<p>Fish for object identification. (<b>a</b>) 100 fish; (<b>b</b>) 150 fish; (<b>c</b>) 200 fish.</p>
Full article ">Figure 18
<p>Trend of number of fish with depth. (<b>a</b>) Comparison bar chart of depth distribution; (<b>b</b>) Comparison line chart of fish depth distribution.</p>
Full article ">Figure 19
<p>Horizontal distribution maps of fish. (<b>a</b>) Horizontal distribution of 100 fish; (<b>b</b>) Horizontal distribution of 150 fish; (<b>c</b>) Horizontal distribution of 200 fish.</p>
Full article ">Figure 20
<p>Horizontal distribution comparison chart of fish.</p>
Full article ">Figure 21
<p>Spatial distribution maps of 100 Fish in each water layer. (<b>a</b>) Water layer: 0–40 cm; (<b>b</b>) Water layer: 25–65 cm; (<b>c</b>) Water layer: 50–90 cm; (<b>d</b>) Water layer: 75–115 cm; (<b>e</b>) Water layer: 100–140 cm; (<b>f</b>) Water layer: 125–165 cm; (<b>g</b>) Water layer: 150–190 cm; (<b>h</b>) Water layer: 175–215 cm; (<b>i</b>) Water layer: 200–240 cm; (<b>j</b>) Water layer: 225–265 cm; (<b>k</b>) Water layer: 250–290 cm.</p>
Full article ">Figure 22
<p>Spatial distribution maps of 150 Fish in each water layer. (<b>a</b>) Water layer: 0–40 cm; (<b>b</b>) Water layer: 25–65 cm; (<b>c</b>) Water layer: 50–90 cm; (<b>d</b>) Water layer: 75–115 cm; (<b>e</b>) Water layer: 100–140 cm; (<b>f</b>) Water layer: 125–165 cm; (<b>g</b>) Water layer: 150–190 cm; (<b>h</b>) Water layer: 175–215 cm; (<b>i</b>) Water layer: 200–240 cm; (<b>j</b>) Water layer: 225–265 cm; (<b>k</b>) Water layer: 250–290 cm.</p>
Full article ">Figure 23
<p>Spatial distribution maps of 200 fish in each water layer. (<b>a</b>) Water layer: 0–40 cm; (<b>b</b>) Water layer: 25–65 cm; (<b>c</b>) Water layer: 50–90 cm; (<b>d</b>) Water layer: 75–115 cm; (<b>e</b>) Water layer: 100–140 cm; (<b>f</b>) Water layer: 125–165 cm; (<b>g</b>) Water layer: 150–190 cm; (<b>h</b>) Water layer: 175–215 cm; (<b>i</b>) Water layer: 200–240 cm; (<b>j</b>) Water layer: 225–265 cm; (<b>k</b>) Water layer: 250–290 cm.</p>
Full article ">Figure 24
<p>3D spatial distribution maps of cage fish. (<b>a</b>) Spatial distribution of 100 fish group; (<b>b</b>) Spatial distribution of 150 fish group; (<b>c</b>) Spatial distribution of 200 fish group.</p>
Full article ">Figure 25
<p>Spatial distribution comparison chart of fish.</p>
Full article ">Figure 26
<p>Distribution of fish at different temperatures. (<b>a</b>) Distribution of 100 fish; (<b>b</b>) Distribution of 150 fish; (<b>c</b>) Distribution of 200 fish.</p>
Full article ">
13 pages, 3604 KiB  
Article
A Super-Resolution and 3D Reconstruction Method Based on OmDF Endoscopic Images
by Fujia Sun and Wenxuan Song
Sensors 2024, 24(15), 4890; https://doi.org/10.3390/s24154890 - 27 Jul 2024
Viewed by 1536
Abstract
In the field of endoscopic imaging, challenges such as low resolution, complex textures, and blurred edges often degrade the quality of 3D reconstructed models. To address these issues, this study introduces an innovative endoscopic image super-resolution and 3D reconstruction technique named Omni-Directional Focus [...] Read more.
In the field of endoscopic imaging, challenges such as low resolution, complex textures, and blurred edges often degrade the quality of 3D reconstructed models. To address these issues, this study introduces an innovative endoscopic image super-resolution and 3D reconstruction technique named Omni-Directional Focus and Scale Resolution (OmDF-SR). This method integrates an Omnidirectional Self-Attention (OSA) mechanism, an Omnidirectional Scale Aggregation Group (OSAG), a Dual-stream Adaptive Focus Mechanism (DAFM), and a Dynamic Edge Adjustment Framework (DEAF) to enhance the accuracy and efficiency of super-resolution processing. Additionally, it employs Structure from Motion (SfM) and Multi-View Stereo (MVS) technologies to achieve high-precision medical 3D models. Experimental results indicate significant improvements in image processing with a PSNR of 38.2902 dB and an SSIM of 0.9746 at a magnification factor of ×2, and a PSNR of 32.1723 dB and an SSIM of 0.9489 at ×4. Furthermore, the method excels in reconstructing detailed 3D models, enhancing point cloud density, mesh quality, and texture mapping richness, thus providing substantial support for clinical diagnosis and surgical planning. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>The structure of Omni Self-Attention (OSA).</p>
Full article ">Figure 2
<p>The structure of Local Convolution Block (LCB).</p>
Full article ">Figure 3
<p>The structure of the Dual-stream Adaptive Focus Mechanism (DAFM).</p>
Full article ">Figure 4
<p>The overall architecture of the proposed OmDF-SR framework.</p>
Full article ">Figure 5
<p>The overall architecture of 3D reconstruction.</p>
Full article ">Figure 6
<p>Training loss of different model.</p>
Full article ">Figure 7
<p>Perceptual results of different models, corresponding HR image, and initial image with enlargement scale factor ×2.</p>
Full article ">Figure 8
<p>Reconstructed 3D model.</p>
Full article ">
22 pages, 7448 KiB  
Article
End-to-End Framework for the Automatic Matching of Omnidirectional Street Images and Building Data and the Creation of 3D Building Models
by Yoshiki Ogawa, Ryoto Nakamura, Go Sato, Hiroya Maeda and Yoshihide Sekimoto
Remote Sens. 2024, 16(11), 1858; https://doi.org/10.3390/rs16111858 - 23 May 2024
Viewed by 1240
Abstract
For accurate urban planning, three-dimensional (3D) building models with a high level of detail (LOD) must be developed. However, most large-scale 3D building models are limited to a low LOD of 1–2, as the creation of higher LOD models requires the modeling of [...] Read more.
For accurate urban planning, three-dimensional (3D) building models with a high level of detail (LOD) must be developed. However, most large-scale 3D building models are limited to a low LOD of 1–2, as the creation of higher LOD models requires the modeling of detailed building elements such as walls, windows, doors, and roof shapes. This process is currently not automated and is performed manually. In this study, an end-to-end framework for the creation of 3D building models was proposed by integrating multi-source data such as omnidirectional images, building footprints, and aerial photographs. These different data sources were matched with the building ID considering their spatial location. The building element information related to the exterior of the building was extracted, and detailed LOD3 3D building models were created. Experiments were conducted using data from Kobe, Japan, yielding a high accuracy for the intermediate processes, such as an 86.9% accuracy in building matching, an 88.3% pixel-based accuracy in the building element extraction, and an 89.7% accuracy in the roof type classification. Eighty-one LOD3 3D building models were created in 8 h, demonstrating that our method can create 3D building models that adequately represent the exterior information of actual buildings. Full article
Show Figures

Figure 1

Figure 1
<p>Five levels of detail in CityGML [Reprinted/adapted with permission from Ref. [<a href="#B9-remotesensing-16-01858" class="html-bibr">9</a>]. 2022, MLIT Japan].</p>
Full article ">Figure 2
<p>Building data used for this study (shooting points of omnidirectional images, building footprints, and aerial photographs). In our study, approximately, 100 m × 100 m of the orange-colored buildings were reconstructed.</p>
Full article ">Figure 3
<p>Examples of texture images.</p>
Full article ">Figure 4
<p>Overview of the proposed framework.</p>
Full article ">Figure 5
<p>The flow of matching omnidirectional images and buildings.</p>
Full article ">Figure 6
<p>Schematic representation of the texture synthesis model.</p>
Full article ">Figure 7
<p>Types of building roofs.</p>
Full article ">Figure 8
<p>Examples of images linked to edges of a building polygon with ID 314960.</p>
Full article ">Figure 9
<p>Results of the semantic segmentation: (<b>a</b>) original building image, (<b>b</b>) ground truth, and (<b>c</b>) prediction.</p>
Full article ">Figure 10
<p>Results of window and door location detection: (<b>a</b>) original building image and (<b>b</b>) prediction.</p>
Full article ">Figure 11
<p>Results of the texture synthesis: (<b>a</b>) original building image, (<b>b</b>) wall crop image, and (<b>c</b>) synthesized texture image.</p>
Full article ">Figure 12
<p>Results of roof color extraction: (<b>a</b>) building polygon cut from an aerial photograph, (<b>b</b>) average color of the polygon, and (<b>c</b>) extracted color using the proposed method.</p>
Full article ">Figure 13
<p>Comparison of the created 3D model with the Google Earth model of the target area.</p>
Full article ">Figure 14
<p>Individual examples of generated 3D models: (<b>a</b>) created 3D model and (<b>b</b>) Google Earth model.</p>
Full article ">
16 pages, 4187 KiB  
Article
An Omnidirectional Image Super-Resolution Method Based on Enhanced SwinIR
by Xiang Yao, Yun Pan and Jingtao Wang
Information 2024, 15(5), 248; https://doi.org/10.3390/info15050248 - 28 Apr 2024
Cited by 1 | Viewed by 1623
Abstract
For the significant distortion problem caused by the special projection method of equi-rectangular projection (ERP) images, this paper proposes an omnidirectional image super-resolution algorithm model based on position information transformation, taking SwinIR as the base. By introducing a space position transformation module that [...] Read more.
For the significant distortion problem caused by the special projection method of equi-rectangular projection (ERP) images, this paper proposes an omnidirectional image super-resolution algorithm model based on position information transformation, taking SwinIR as the base. By introducing a space position transformation module that supports deformable convolution, the image preprocessing process is optimized to reduce the distortion effects in the polar regions of the ERP image. Meanwhile, by introducing deformable convolution in the deep feature extraction process, the model’s adaptability to local deformations of images is enhanced. Experimental results on publicly available datasets have shown that our method outperforms SwinIR, with an average improvement of over 0.2 dB in WS-PSNR and over 0.030 in WS-SSIM for ×4 pixel upscaling. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>The Network Architecture of Our Proposed Model (This model aims to convert low resolution (LR) images into high resolution (HR) images).</p>
Full article ">Figure 2
<p>LTM Module.</p>
Full article ">Figure 3
<p>Affine Transformation Diagram.</p>
Full article ">Figure 4
<p>The Process of Deformable Convolution.</p>
Full article ">Figure 5
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 6
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 7
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 8
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 9
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 10
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 11
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 12
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 13
<p>Qualitative Comparison of ×4 Pixel Upsampling Results.</p>
Full article ">Figure 14
<p>Changes in Results as Training Progresses.</p>
Full article ">Figure 15
<p>Changes in Loss Function.</p>
Full article ">
22 pages, 7518 KiB  
Article
Omni-OTPE: Omnidirectional Optimal Real-Time Ground Target Position Estimation System for Moving Lightweight Unmanned Aerial Vehicle
by Yi Ding, Jiaxing Che, Zhiming Zhou and Jingyuan Bian
Sensors 2024, 24(5), 1709; https://doi.org/10.3390/s24051709 - 6 Mar 2024
Viewed by 1399
Abstract
Ground target detection and positioning systems based on lightweight unmanned aerial vehicles (UAVs) are increasing in value for aerial reconnaissance and surveillance. However, the current method for estimating the target’s position is limited by the field of view angle, rendering it challenging to [...] Read more.
Ground target detection and positioning systems based on lightweight unmanned aerial vehicles (UAVs) are increasing in value for aerial reconnaissance and surveillance. However, the current method for estimating the target’s position is limited by the field of view angle, rendering it challenging to fulfill the demands of a real-time omnidirectional reconnaissance operation. To address this issue, we propose an Omnidirectional Optimal Real-Time Ground Target Position Estimation System (Omni-OTPE) that utilizes a fisheye camera and LiDAR sensors. The object of interest is first identified in the fisheye image, and then, the image-based target position is obtained by solving using the fisheye projection model and the target center extraction algorithm based on the detected edge information. Next, the LiDAR’s real-time point cloud data are filtered based on position–direction constraints using the image-based target position information. This step allows for the determination of point cloud clusters that are relevant to the characterization of the target’s position information. Finally, the target positions obtained from the two methods are fused using an optimal Kalman fuser to obtain the optimal target position information. In order to evaluate the positioning accuracy, we designed a hardware and software setup, mounted on a lightweight UAV, and tested it in a real scenario. The experimental results validate that our method exhibits significant advantages over traditional methods and achieves a real-time high-performance ground target position estimation function. Full article
Show Figures

Figure 1

Figure 1
<p>Omni-OTPE: Omnidirectional optimal real-time ground target position estimation system framework.</p>
Full article ">Figure 2
<p>The drone is equipped with Omni-OTPE system that allows it to estimate the position of a target-tagged object in conjunction with its own position.</p>
Full article ">Figure 3
<p>The layout of the sensors and the range of vertical sensing angles.</p>
Full article ">Figure 4
<p>Yolov5 performance plots for each model.</p>
Full article ">Figure 5
<p>Comparison of distortion of different projection models. The model includes Keyhole projection, Stereo graphic projection, Equidistance projection, Equisolid projection and Orthogonal projection. Pinhole projection is the standard camera model, which theoretically has no distortion. The degree of distortion of the model can be expressed as the degree of curve deviation between this model and pinhole projection model.</p>
Full article ">Figure 6
<p>Description of the center offset problem.</p>
Full article ">Figure 7
<p>(<b>a</b>) Polar and azimuthal angles corresponding to the upper left (red) and lower right (green) fixed points of the identification box in the fisheye image. (<b>b</b>) Fisheye camera point projection process.</p>
Full article ">Figure 8
<p>(<b>a</b>) Position-based point cloud filtering. (<b>b</b>) Direction-based point cloud filtering, where the green dots are the point cloud generated by the target; the red dots indicate anomalies; and the blue dots indicate vision-based position estimates <math display="inline"><semantics> <mrow> <mmultiscripts> <mi>P</mi> <mi mathvariant="italic">tag</mi> <mi>V</mi> <mprescripts/> <none/> <mi>w</mi> </mmultiscripts> </mrow> </semantics></math>.</p>
Full article ">Figure 9
<p>Time series of the target location output.</p>
Full article ">Figure 10
<p>Optimal target position information fusion process.</p>
Full article ">Figure 11
<p>Scene detection graph. There are three box objects of similar size in the scene. The one with the gray surface and label is the target.</p>
Full article ">Figure 12
<p>(<b>a</b>) Flight experiment in which there are two distractors and one target object; (<b>b</b>) UAV trajectory and target position estimates, where the color of the trajectory can indicate the distance from the UAV to the target.</p>
Full article ">Figure 13
<p>Image-based plot of horizontal coordinates of the estimated points of the target position, indicating the distribution of the estimated points. (<b>a</b>) Estimated target position at 5 m altitude by the UAV. where the ellipse denotes the 95 confidence ellipse, which can indicate the degree of distribution of the position estimates. (<b>b</b>) Estimated target position at 7 m altitude by the UAV.</p>
Full article ">Figure 14
<p>Direction angle error curves for both methods. (<b>a</b>) Angular error in position estimation in flight at 5 m altitude. (<b>b</b>) Angular error in position estimation in flight at 7 m altitude.</p>
Full article ">Figure 15
<p>Target position estimation error curves for both methods. (<b>a</b>) Positioning error in position estimation in flight at 5 m altitude. (<b>b</b>) Positioning error in position estimation in flight at 7 m altitude.</p>
Full article ">Figure 16
<p>When the visual localization error is too large, the back-end will reject the correct point cloud information, reducing the back-end repositioning performance.</p>
Full article ">Figure 17
<p>Three positional estimation error curves. (<b>a</b>) Target localization error of each localization method in 5 m flight experiment. (<b>b</b>) Target localization error of each localization method in 7 m flight experiment. The red point is the error of LiDAR localization, and the red area indicates that the system finds the corresponding target point cloud cluster at that moment based on the visual localization result. The blue curve is the position error curve after fusion.</p>
Full article ">
21 pages, 1865 KiB  
Article
Latitude-Adaptive Integer Bit Allocation for Quantization of Omnidirectional Images
by Qian Sima, Hui Feng and Bo Hu
Appl. Sci. 2024, 14(5), 1861; https://doi.org/10.3390/app14051861 - 23 Feb 2024
Cited by 1 | Viewed by 1134
Abstract
Omnidirectional images have gained significant popularity and drawn great attention nowadays, which poses challenges to omnidirectional image processing in solving the bottleneck of storage and transmission. Projecting onto a two-dimensional image plane is generally used to compress an omnidirectional image. However, the most [...] Read more.
Omnidirectional images have gained significant popularity and drawn great attention nowadays, which poses challenges to omnidirectional image processing in solving the bottleneck of storage and transmission. Projecting onto a two-dimensional image plane is generally used to compress an omnidirectional image. However, the most commonly used projection format, the equirectangular projection (ERP), results in a significant amount of redundant samples in the polar areas, thus incurring extra bitrate and geometric distortion. We derive the optimal latitude-adaptive bit allocation for each image tile. Subsequently, we propose a greedy algorithm for non-negative integer bit allocation (NNIBA) for non-uniform quantization under an omnidirectional image quality metric WMSE. In our experiment, we design quantization tables based on JPEG and compare our approach with other sampling-related methods. Our method achieves an average bit saving of 7.9% compared with JPEG while outperforming other sampling-related methods. Besides, we compare our non-uniform quantization approach with two proposed bit allocation methods, achieving an average improvement of 0.35 dB and 2.66 dB under WS-PSNR, respectively. The visual quality assessment also confirms the superiority of our method. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of omnidirectional image compression methods.</p>
Full article ">Figure 2
<p>Processing chain of our method. (<b>a</b>) A batch of original images. (<b>b</b>) Coefficient matrices partitioned into blocks. (<b>c</b>) Recovered images after quantization. (<b>d</b>) Distribution modeling of coefficients at different latitudes. (<b>e</b>) Bit allocation among titles. (<b>f</b>) Subsequent bit allocation in each tile.</p>
Full article ">Figure 3
<p>Coefficient frequencies at latitude 1. (<b>a</b>) DC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>1</mn> </mrow> </msub> </semantics></math>. (<b>b</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> </msub> </semantics></math>. (<b>c</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>30</mn> </mrow> </msub> </semantics></math>. (<b>d</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>1</mn> <mo>,</mo> <mn>64</mn> </mrow> </msub> </semantics></math>.</p>
Full article ">Figure 4
<p>Coefficients frequencies at latitude 17. (<b>a</b>) DC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>17</mn> <mo>,</mo> <mn>1</mn> </mrow> </msub> </semantics></math>. (<b>b</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>17</mn> <mo>,</mo> <mn>2</mn> </mrow> </msub> </semantics></math>. (<b>c</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>17</mn> <mo>,</mo> <mn>30</mn> </mrow> </msub> </semantics></math>. (<b>d</b>) AC coefficients <math display="inline"><semantics> <msub> <mi mathvariant="bold">c</mi> <mrow> <mn>17</mn> <mo>,</mo> <mn>64</mn> </mrow> </msub> </semantics></math>.</p>
Full article ">Figure 5
<p>Rate distortion curves for JPEG-based methods over 100 images with a resolution of <math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>1024</mn> </mrow> </semantics></math> pixels [<a href="#B10-applsci-14-01861" class="html-bibr">10</a>,<a href="#B20-applsci-14-01861" class="html-bibr">20</a>,<a href="#B23-applsci-14-01861" class="html-bibr">23</a>].</p>
Full article ">Figure 6
<p>Rate distortion curves for non-negative integer bit allocation (NNIBA) methods over 100 images with a resolution of <math display="inline"><semantics> <mrow> <mn>256</mn> <mo>×</mo> <mn>512</mn> </mrow> </semantics></math> pixels [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 7
<p>Rate distortion curves for non-negative integer bit allocation (NNIBA) methods over 100 images with a resolution of <math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>1024</mn> </mrow> </semantics></math> pixels [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 8
<p>Four test images from Pano 3D dataset with a resolution of <math display="inline"><semantics> <mrow> <mn>256</mn> <mo>×</mo> <mn>512</mn> </mrow> </semantics></math>. The red box is used for subsequent zoomed comparisons.</p>
Full article ">Figure 9
<p>Zoomed version of test image A at 0.4 bpp [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 10
<p>Zoomed version of test image B at 0.4 bpp [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 11
<p>Zoomedversion of test image C at 0.75 bpp [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">Figure 12
<p>Zoomedversion of test image D at 0.75 bpp [<a href="#B25-applsci-14-01861" class="html-bibr">25</a>,<a href="#B31-applsci-14-01861" class="html-bibr">31</a>].</p>
Full article ">
18 pages, 10954 KiB  
Article
Using a Robot for Indoor Navigation and Door Opening Control Based on Image Processing
by Chun-Hsiang Hsu and Jih-Gau Juang
Actuators 2024, 13(2), 78; https://doi.org/10.3390/act13020078 - 16 Feb 2024
Viewed by 1880
Abstract
This study used real-time image processing to realize obstacle avoidance and indoor navigation with an omnidirectional wheeled mobile robot (WMR). The distance between an obstacle and the WMR was obtained using a depth camera. Real-time images were used to control the robot’s movements. [...] Read more.
This study used real-time image processing to realize obstacle avoidance and indoor navigation with an omnidirectional wheeled mobile robot (WMR). The distance between an obstacle and the WMR was obtained using a depth camera. Real-time images were used to control the robot’s movements. The WMR can extract obstacle distance data from a depth map and apply fuzzy theory to avoid obstacles in indoor environments. A fuzzy control system was integrated into the control scheme. After detecting a doorknob, the robot could track the target and open the door. We used the speeded up robust features matching algorithm to recognize the WMR’s movement direction. The proposed control scheme ensures that the WMR can avoid obstacles, move to a designated location, and open a door. Like humans, the robot performs the described task only using visual sensors. Full article
(This article belongs to the Special Issue Actuators in Robotic Control—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>The primary devices of the proposed robot system are (<b>a</b>) a mobile robot; (<b>b</b>) an omnidirectional wheel, battery, and control components; and (<b>c</b>) an Intel Realsense depth camera.</p>
Full article ">Figure 2
<p>(<b>a</b>) Omnidirectional wheel structure. (<b>b</b>) Coordinate system.</p>
Full article ">Figure 3
<p>(<b>a</b>) Arduino Uno R3 [<a href="#B15-actuators-13-00078" class="html-bibr">15</a>], (<b>b</b>) DFRduino IO Expansion board, and (<b>c</b>) Microsoft LifeCam.</p>
Full article ">Figure 4
<p>Zhang Zhengyou camera calibration samples.</p>
Full article ">Figure 5
<p>Stereo rectification.</p>
Full article ">Figure 6
<p>Mug distance.</p>
Full article ">Figure 7
<p>(<b>a</b>) is the original image, (<b>b</b>) is the depth image where the blue color means the object is near to the camera, and dark red means the object is far from the camera. The distance ranges from 0 m (dark blue) to 8 m (dark red).</p>
Full article ">Figure 8
<p>Detection of different obstacles: (<b>a</b>) obstacle detection and obstacle depth map, (<b>b</b>) nearest obstacle point is on the right side, (<b>c</b>) move to the left and the obstacle is outside the safe frame, (<b>d</b>) nearest obstacle point is on the left side, (<b>e</b>) move to the right and the moving direction is clear.</p>
Full article ">Figure 9
<p>SURF matching: (<b>a</b>) robot is heading in the right direction, (<b>b</b>) robot is heading in the wrong direction, (<b>c</b>) robot is heading in the right direction.</p>
Full article ">Figure 10
<p>Edge detection uses a threshold, and the doorknob is detected; the Chinese character on the left part of the figure is the room’s name.</p>
Full article ">Figure 11
<p>Control flowchart.</p>
Full article ">Figure 12
<p>Robot moving path.</p>
Full article ">Figure 13
<p>The fuzzy sets of the three inputs are near (blue), medium (red), and far (yellow); the fuzzy sets of the output pixel are turn_left (blue), go_straight (red), and turn_right (yellow); the fuzzy sets of the output time are short (blue), medium (red), and long (yellow).</p>
Full article ">Figure 14
<p>Fuzzy control model where the yellow color means large value and the dark blue means small value.</p>
Full article ">Figure 15
<p>Fuzzy control scheme.</p>
Full article ">Figure 16
<p>The robot arm lifts to the specified height; the Chinese characters on the figure are the room’s name.</p>
Full article ">Figure 17
<p>Robot obstacle avoidance test.</p>
Full article ">Figure 18
<p>On the depth camera image, the distances of the nearest left, middle, and right objects are greater than 2 m, so the action is GO.</p>
Full article ">Figure 19
<p>Robot’s starting position.</p>
Full article ">Figure 20
<p>The robot finds an obstacle as the blue mark (in the middle area) and yellow mark (in the right area) on the picture, and the distances are 1.65 m and 1.62 m, respectively; the action is a LEFT turn.</p>
Full article ">Figure 21
<p>Robot avoids obstacle (box).</p>
Full article ">Figure 22
<p>Image on the robot’s depth camera.</p>
Full article ">Figure 23
<p>(<b>a</b>) The robot reaches the midway point; (<b>b</b>) the robot’s image.</p>
Full article ">Figure 24
<p>After the SURF matching, the robot turns to the target direction and moves forward.</p>
Full article ">Figure 25
<p>(<b>a</b>) The robot arrives at the specified location; (<b>b</b>) the Chinese characters on the figure are the room’s name.</p>
Full article ">Figure 26
<p>(<b>a</b>) Doorknob detection, Activate arm and arm’s camera; (<b>b</b>) the Chinese characters on the figure are the room’s name. Arm’s camera image; the object distance is 0.827 m, and the coordinate is (436.5, 51.5).</p>
Full article ">Figure 27
<p>The robot opens the door; the Chinese characters on the figure are the room’s name.</p>
Full article ">
16 pages, 9434 KiB  
Article
Omnidirectional-Sensor-System-Based Texture Noise Correction in Large-Scale 3D Reconstruction
by Wenya Xie and Xiaoping Hong
Sensors 2024, 24(1), 78; https://doi.org/10.3390/s24010078 - 22 Dec 2023
Viewed by 1238
Abstract
The evolution of cameras and LiDAR has propelled the techniques and applications of three-dimensional (3D) reconstruction. However, due to inherent sensor limitations and environmental interference, the reconstruction process often entails significant texture noise, such as specular highlight, color inconsistency, and object occlusion. Traditional [...] Read more.
The evolution of cameras and LiDAR has propelled the techniques and applications of three-dimensional (3D) reconstruction. However, due to inherent sensor limitations and environmental interference, the reconstruction process often entails significant texture noise, such as specular highlight, color inconsistency, and object occlusion. Traditional methodologies grapple to mitigate such noise, particularly in large-scale scenes, due to the voluminous data produced by imaging sensors. In response, this paper introduces an omnidirectional-sensor-system-based texture noise correction framework for large-scale scenes, which consists of three parts. Initially, we obtain a colored point cloud with luminance value through LiDAR points and RGB images organization. Next, we apply a voxel hashing algorithm during the geometry reconstruction to accelerate the computation speed and save the computer memory. Finally, we propose the key innovation of our paper, the frame-voting rendering and the neighbor-aided rendering mechanisms, which effectively eliminates the aforementioned texture noise. From the experimental results, the processing rate of one million points per second shows its real-time applicability, and the output figures of texture optimization exhibit a significant reduction in texture noise. These results indicate that our framework has advanced performance in correcting multiple texture noise in large-scale 3D reconstruction. Full article
(This article belongs to the Special Issue Sensing and Processing for 3D Computer Vision: 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Specular highlight phenomenon. (<b>b</b>) The position of the highlight areas in the image changes with the variation of the sensor pose. In the image, the red box indicates the most prominent highlight noise, and the green box indicates the door, which serves as a positional reference.</p>
Full article ">Figure 2
<p>Color inconsistency phenomenon. P1–P3 are three consecutive images in terms of position. (<b>a</b>) Normal situation with consistent color between frames. (<b>b</b>) Inconsistent color between frames caused by variations in the intensity of the light source or changes in its relative position to the sensor.</p>
Full article ">Figure 3
<p>Pipeline of the whole process, consisting of data organization, geometry reconstruction, and texture optimization.</p>
Full article ">Figure 4
<p>Process flow of data organization. (<b>a</b>) RGB image. (<b>b</b>) CIELAB color space image transformed from RGB image, which facilitates luminance evaluation in the subsequent section of our work. (<b>c</b>) LiDAR point cloud. (<b>d</b>) Fusion of LiDAR point cloud with RGB image. (<b>e</b>) Fusion of LiDAR point cloud with CIELAB color space image.</p>
Full article ">Figure 5
<p>Voxel hashing schematic. The mapping between point coordinates and voxel block indices is achieved through a hash table, thereby efficiently allocating points while making reasonable use of computer storage resources.</p>
Full article ">Figure 6
<p>Motivation for proposing neighbor-aided rendering mechanism: points are randomly distributed in voxels; thus, some voxels lack insufficient points for self-optimization.</p>
Full article ">Figure 7
<p>Neighbor-aided rendering mechanism. The figure illustrates the configuration of a voxel block and the interconnections between adjacent voxels.</p>
Full article ">Figure 8
<p>Sensor setup for data collection.</p>
Full article ">Figure 9
<p>Input data. The dataset consists of four spots, and each spot consists of five specified poses.</p>
Full article ">Figure 10
<p>Highlight noise correction in scene 1 according to frame-voting rendering. Regions (<b>a</b>)–(<b>c</b>) present specular highlights phenomenon on the screen and wall surfaces in the scene.</p>
Full article ">Figure 11
<p>Elimination of object occlusion in scene 2 with frame-voting rendering. (<b>a</b>) Comparison diagram of the elimination of misimaging caused by table occlusion. (<b>b</b>) Comparison diagram of the elimination of misimaging caused by chair occlusion.</p>
Full article ">Figure 12
<p>Enhanced outcome with neighbor-aided optimization. Regions A–C exhibite pronounced contrastive effects. (<b>a</b>) Demonstration area of the original point cloud containing numerous types of texture noise. (<b>b</b>) The result optimized using only frame-voting rendering. (<b>c</b>) The result optimized further with neighbor-aided rendering.</p>
Full article ">Figure 13
<p>Comparing results of highlight removal method. (<b>a</b>) Projection of raw model (input). The white boxes indicate areas with noise that should be corrected. The red box indicates area that should not be corrected (lights). (<b>b</b>) Projection of texture optimized model (ours). (<b>c</b>) Yang et al. (2010) [<a href="#B2-sensors-24-00078" class="html-bibr">2</a>]. (<b>d</b>) Shen et al. (2013) [<a href="#B3-sensors-24-00078" class="html-bibr">3</a>]. (<b>e</b>) Fu et al. (2019) [<a href="#B4-sensors-24-00078" class="html-bibr">4</a>]. (<b>f</b>) Jin et al. (2023) [<a href="#B8-sensors-24-00078" class="html-bibr">8</a>].</p>
Full article ">
21 pages, 54036 KiB  
Article
Plant Foliar Geometry as a Biomimetic Template for Antenna Design
by Jose Ignacio Lozano, Marco A. Panduro, Rodrigo Méndez-Alonzo, Miguel A. Alonso-Arevalo, Roberto Conte and Alberto Reyna
Biomimetics 2023, 8(7), 531; https://doi.org/10.3390/biomimetics8070531 - 7 Nov 2023
Cited by 2 | Viewed by 1991
Abstract
Plant diversity includes over 300,000 species, and leaf structure is one of the main targets of selection, being highly variable in shape and size. On the other hand, the optimization of antenna design has no unique solution to satisfy the current range of [...] Read more.
Plant diversity includes over 300,000 species, and leaf structure is one of the main targets of selection, being highly variable in shape and size. On the other hand, the optimization of antenna design has no unique solution to satisfy the current range of applications. We analyzed the foliar geometries of 100 plant species and applied them as a biomimetic design template for microstrip patch antenna systems. From this set, a subset of seven species were further analyzed, including species from tropical and temperate forests across the phylogeny of the Angiosperms. Foliar geometry per species was processed by image processing analyses, and the resultant geometries were used in simulations of the reflection coefficients and the radiation patterns via finite differences methods. A value below −10 dB is set for the reflection coefficient to determine the operation frequencies of all antenna elements. All species showed between 3 and 15 operational frequencies, and four species had operational frequencies that included the 2.4 and 5 GHz bands. The reflection coefficients and the radiation patterns in most of the designs were equal or superior to those of conventional antennas, with several species showing multiband effects and omnidirectional radiation. We demonstrate that plant structures can be used as a biomimetic tool in designing microstrip antenna for a wide range of applications. Full article
(This article belongs to the Special Issue Bio-Inspired Design: Creativity and Innovation)
Show Figures

Figure 1

Figure 1
<p>Basic method to assess the structure or each leaf design as a radiating element.</p>
Full article ">Figure 2
<p><span class="html-italic">Tetrapterys macrocarpa</span> Malpighiaceae (elliptic) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> parameter, (<b>c</b>) 3D pattern at 0.87 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 0.87 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 3.3 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 2 Cont.
<p><span class="html-italic">Tetrapterys macrocarpa</span> Malpighiaceae (elliptic) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> parameter, (<b>c</b>) 3D pattern at 0.87 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 0.87 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 3.3 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 3
<p><span class="html-italic">Sarcorhachis naranjoana</span> Piperaceae (ovate) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> parameter, (<b>c</b>) 3D pattern at 1.05 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 1.05 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 3.09 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 4
<p><span class="html-italic">Cissampelos owariensis</span> Menispermaceae (circular base) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D pattern at 1.2 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 1.2 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 6.27 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 4 Cont.
<p><span class="html-italic">Cissampelos owariensis</span> Menispermaceae (circular base) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D pattern at 1.2 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 1.2 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 6.27 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 5
<p><span class="html-italic">Paranomus spectrum</span> Proteaceae (obovate) (<b>a</b>) Front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D radiation pattern at 2.34 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 2.34 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 5.31 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 6
<p><span class="html-italic">Cesearia ilicifolia</span> Salicaceae (toothed) (<b>a</b>) Front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D radiation pattern at 5.91 GHz, and cuts of the radiation pattern at <span class="html-italic">f</span> = 5.91 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 6.51 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 7
<p><span class="html-italic">Liquidambar styraciflua</span> Hamamelidaceae (lobed leaves) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D pattern at 4.23 GHz, and cuts of the pattern at <span class="html-italic">f</span> = 4.23 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 5.58 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 7 Cont.
<p><span class="html-italic">Liquidambar styraciflua</span> Hamamelidaceae (lobed leaves) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D pattern at 4.23 GHz, and cuts of the pattern at <span class="html-italic">f</span> = 4.23 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 5.58 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">Figure 8
<p><span class="html-italic">Quercus alba</span> Fagaceae (pinnately lobed) (<b>a</b>) front view, (<b>b</b>) <span class="html-italic">S</span><sub>11</sub> Parameter, (<b>c</b>) 3D radiation pattern at 3.18 GHz, and cuts of the pattern at <span class="html-italic">f</span> = 3.18 GHz (<b>d</b>) vertical and (<b>e</b>) horizontal cut, and at <span class="html-italic">f</span> = 5.88 GHz (<b>f</b>) vertical and (<b>g</b>) horizontal cut.</p>
Full article ">
19 pages, 24057 KiB  
Article
A Four Element Stringray-Shaped MIMO Antenna System for UWB Applications
by Hüseyin Şerif Savcı
Micromachines 2023, 14(10), 1944; https://doi.org/10.3390/mi14101944 - 18 Oct 2023
Cited by 7 | Viewed by 1661
Abstract
This paper presents a CoPlanar-Waveguide (CPW)-fed stingray-shaped Ultra-WideBand (UWB) Multiple-Input–Multiple-Output (MIMO) antenna system designed for microwave imaging applications. Featuring a diagonal square with four inner lines and a vertical line at the center from toe to tip with a CPW feed line, the [...] Read more.
This paper presents a CoPlanar-Waveguide (CPW)-fed stingray-shaped Ultra-WideBand (UWB) Multiple-Input–Multiple-Output (MIMO) antenna system designed for microwave imaging applications. Featuring a diagonal square with four inner lines and a vertical line at the center from toe to tip with a CPW feed line, the unit antenna element looks like a stingray fish skeleton and is, therefore, named as a stingray-shaped antenna. It offers a bandwidth spanning from 3.8 to 12.7 GHz. Fabricated on a 31mil RO5880 RF teflon substrate with a relative permittivity of 2.2, the proposed antenna has dimensions of 26 × 29 × 0.787 mm3. The maximum realized gain achieved is 3.5 dBi with stable omnidirectional radiation patterns. The antenna element is used in a four-antenna MIMO configuration with an isolation-improving structure at the center. The MIMO system has dimensions of 58 × 58 × 0.787 mm3 with a maximum realized gain of 5.3 dBi. The antenna’s performance in terms of MIMO parameters like Envelope Correlation Coefficient (ECC) and Diversity Gain (DG) is within satisfactory limits for medical imaging applications. Time domain analysis also yields positive results, allowing its integration into a breast phantom tumor detection simulation. The simulation and measurement results demonstrate excellent agreement, making this antenna a promising candidate for microwave imaging and biomedical applications. Full article
(This article belongs to the Special Issue Advances in Microwave/Millimeter-Wave Devices and Antennas)
Show Figures

Figure 1

Figure 1
<p>Near-field microwave imaging setup with a tumor inside beast phantom (<b>a</b>) with twelve single antenna elements, (<b>b</b>) with four MIMO antenna systems.</p>
Full article ">Figure 2
<p>Proposed CPW-fed UWB Antenna.</p>
Full article ">Figure 3
<p>Design evolution (<b>a</b>) stage 1, (<b>b</b>) stage 2, (<b>c</b>) proposed, (<b>d</b>) reflection coefficient.</p>
Full article ">Figure 4
<p>Single antenna element (<b>a</b>) parametric analysis of ground arc length, <math display="inline"><semantics> <msub> <mi>R</mi> <mi>r</mi> </msub> </semantics></math> (<b>b</b>) parametric analysis of feedline length, <math display="inline"><semantics> <msub> <mi>L</mi> <mi>f</mi> </msub> </semantics></math> (<b>c</b>) radiation efficiency, total efficiency and gain over frequency (<b>d</b>) reflection coefficient in dB (<b>e</b>) prototype photo (<b>f</b>) s-parameter measurement setup.</p>
Full article ">Figure 5
<p>Time-domain response of single antenna element (<b>a</b>) face-to-face configuration (<b>b</b>) side-by-side configuration (<b>c</b>) face-to-face response (<b>d</b>) side-by-side response.</p>
Full article ">Figure 6
<p>Group delay of UWB antenna.</p>
Full article ">Figure 7
<p>Four element MIMO antennas (<b>a</b>) simple configuration (<b>b</b>) with isolating structure at the center.</p>
Full article ">Figure 8
<p>The scattering parameters of four element MIMO antennas (<b>a</b>,<b>b</b>) simple configuration (<b>c</b>,<b>d</b>) with isolating structure at the center.</p>
Full article ">Figure 8 Cont.
<p>The scattering parameters of four element MIMO antennas (<b>a</b>,<b>b</b>) simple configuration (<b>c</b>,<b>d</b>) with isolating structure at the center.</p>
Full article ">Figure 9
<p>The surface current density for the four elements MIMO antennas (<b>a</b>) at 4.5 GHz without isolating element (<b>b</b>) at 4.5 GHz with isolating element (<b>c</b>) at 6.7 GHz without isolating element (<b>d</b>) at 6.7 GHz with isolating element (<b>e</b>) at 15.5 GHz without isolating element (<b>f</b>) at 15.5 GHz with isolating element.</p>
Full article ">Figure 10
<p>Four-element MIMO antenna system with isolating structure at the center.</p>
Full article ">Figure 11
<p>(<b>a</b>) Measured reflection coefficient, (<b>b</b>) Measured port-to-port isolation.</p>
Full article ">Figure 12
<p>2D Radiation Patterns (<b>a</b>) 6.7 GHz Ant 1, (<b>b</b>) 6.7 GHz Ant 2, (<b>c</b>) 6.7 GHz Ant 3, (<b>d</b>) 6.7 GHz Ant 4, (<b>e</b>) 11 GHz Ant 1, (<b>f</b>) 11 GHz Ant 2, (<b>g</b>) 11 GHz Ant 3, (<b>h</b>) 11 GHz Ant 4.</p>
Full article ">Figure 12 Cont.
<p>2D Radiation Patterns (<b>a</b>) 6.7 GHz Ant 1, (<b>b</b>) 6.7 GHz Ant 2, (<b>c</b>) 6.7 GHz Ant 3, (<b>d</b>) 6.7 GHz Ant 4, (<b>e</b>) 11 GHz Ant 1, (<b>f</b>) 11 GHz Ant 2, (<b>g</b>) 11 GHz Ant 3, (<b>h</b>) 11 GHz Ant 4.</p>
Full article ">Figure 13
<p>MIMO parameters: (<b>a</b>) ECC simulation (with ISO), (<b>b</b>) ECC measurement (with ISO), (<b>c</b>) DG (with ISO), (<b>d</b>) ECC (without ISO), (<b>e</b>) CC (with ISO).</p>
Full article ">Figure 13 Cont.
<p>MIMO parameters: (<b>a</b>) ECC simulation (with ISO), (<b>b</b>) ECC measurement (with ISO), (<b>c</b>) DG (with ISO), (<b>d</b>) ECC (without ISO), (<b>e</b>) CC (with ISO).</p>
Full article ">
Back to TopTop