[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,139)

Search Parameters:
Keywords = low-cost camera

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
15 pages, 4538 KiB  
Article
Fish Detection in Fishways for Hydropower Stations Using Bidirectional Cross-Scale Feature Fusion
by Junming Wang, Yuanfeng Gong, Wupeng Deng, Enshun Lu, Xinyu Hu and Daode Zhang
Appl. Sci. 2025, 15(5), 2743; https://doi.org/10.3390/app15052743 - 4 Mar 2025
Viewed by 206
Abstract
Fishways can effectively validate the effectiveness and rationality of their construction, optimize operational modes, and achieve intelligent scientific management through fish species detection. Traditional fish species detection methods for fishways are unsuitable due to inefficiency and disruption of the fish ecological environment. Therefore, [...] Read more.
Fishways can effectively validate the effectiveness and rationality of their construction, optimize operational modes, and achieve intelligent scientific management through fish species detection. Traditional fish species detection methods for fishways are unsuitable due to inefficiency and disruption of the fish ecological environment. Therefore, combining cameras with target detection technology provides a better solution. However, challenges include the limited computational power of onsite equipment, the complexity of model deployment, low detection accuracy, and slow detection speed, all of which are significant obstacles. This paper proposes a fish detection model for accurate and efficient fish detection. Firstly, the backbone network integrates FasterNet-Block, C2f, and an efficient multi-scale EMA attention mechanism to address attention dispersion problems during feature extraction, delivering real-time object detection across different scales. Secondly, the Neck introduces a novel architecture to enhance feature fusion by integrating the RepBlock and BiFusion modules. Finally, the performance of the fish detection model is demonstrated based on the Fish26 dataset, in which the detection accuracy, computational cost, and parameter count are significantly optimized by 1.7%, 23.4%, and 24%, respectively, compared to the state-of-the-art model. At the same time, we installed detection devices in a specific fishway and deployed the proposed method within these devices. We collected data on four fish species passing through the fishway to create a dataset and train the model. The results of the practical application demonstrated superior fish detection capabilities, with rapid detection ability achieved while minimizing resource usage. This validated the effectiveness of the proposed method for equipment deployment in real-world engineering environments. This marks a shift from traditional manual detection to intelligent fish species detection in fishways, promoting water resource utilization and the protection of fish ecological environments. Full article
Show Figures

Figure 1

Figure 1
<p>The environment around the fishway and the underwater imaging of the fishway.</p>
Full article ">Figure 2
<p>The overall framework of the fishway fish detection model and the details of some modules.</p>
Full article ">Figure 3
<p>The improved structure diagram of each part of the C2f module: (<b>a</b>) Structure diagram of the Faster Block module and a schematic of the convolution principle. (<b>b</b>) Structure diagram of the C2f-Faster-EMA module after the improvement of the C2f module using the Faster Block and the addition of the EMA attention mechanism. (<b>c</b>) Refined flowchart of the EMA attention mechanism module.</p>
Full article ">Figure 4
<p>The overall flowchart of the neck part in the fishway fish detection and the module operation process at each stage: (<b>a</b>) Overall flow structure diagram of RepBi-PAN; (<b>b</b>) Operation flowchart of the RepBlock module in the training phase of the RepBi-PAN structure; (<b>c</b>) Operation flowchart of the RepBlock module in the inference phase of the RepBi-PAN structure; (<b>d</b>) Structure diagram of the BiFusion module in the RepBi-PAN process.</p>
Full article ">Figure 5
<p>Sample Images from the Fish26 Dataset.</p>
Full article ">Figure 6
<p>The mAP@0.5 value curve of different detection algorithms after training and testing on the Fish26 dataset.</p>
Full article ">Figure 7
<p>The performance of the proposed fishway fish detection model during the training and validation process on the Fish26 dataset, including various metrics.</p>
Full article ">Figure 8
<p>The 3D schematic of the underwater detection system, consisting of a fish passage box culvert and an encapsulated camera, and its installation at the designated location at the fishway inlet after processing and assembly.</p>
Full article ">Figure 9
<p>The fishway detection system was installed at the designated position of the fishway outlet using an electric hoist, and underwater images of different fish species in the fishway were captured.</p>
Full article ">Figure 10
<p>The proposed fishway fish detection model can perform real-time monitoring of different fish species in real fishways.</p>
Full article ">
20 pages, 6141 KiB  
Article
Development of Low-Cost Monitoring and Assessment System for Cycle Paths Based on Raspberry Pi Technology
by Salvatore Bruno, Ionut Daniel Trifan, Lorenzo Vita and Giuseppe Loprencipe
Infrastructures 2025, 10(3), 50; https://doi.org/10.3390/infrastructures10030050 - 2 Mar 2025
Viewed by 237
Abstract
Promoting alternative modes of transportation such as cycling represents a valuable strategy to minimize environmental impacts, as confirmed in the main targets set out by the European Commission. In this regard, in cities throughout the world, there has been a significant increase in [...] Read more.
Promoting alternative modes of transportation such as cycling represents a valuable strategy to minimize environmental impacts, as confirmed in the main targets set out by the European Commission. In this regard, in cities throughout the world, there has been a significant increase in the construction of bicycle paths in recent years, requiring effective maintenance strategies to preserve their service levels. The continuous monitoring of road networks is required to ensure the timely scheduling of optimal maintenance activities. This involves regular inspections of the road surface, but there are currently no automated systems for monitoring cycle paths. In this study, an integrated monitoring and assessment system for cycle paths was developed exploiting Raspberry Pi technologies. In more detail, a low-cost Inertial Measurement Unit (IMU), a Global Positioning System (GPS) module, a magnetic Hall Effect sensor, a camera module, and an ultrasonic distance sensor were connected to a Raspberry Pi 4 Model B. The novel system was mounted on a e-bike as a test vehicle to monitor the road conditions of various sections of cycle paths in Rome, characterized by different pavement types and decay levels as detected using the whole-body vibration awz index (ISO 2631 standard). Repeated testing confirmed the system’s reliability by assigning the same vibration comfort class in 74% of the cases and an adjacent one in 26%, with an average difference of 0.25 m/s2, underscoring its stability and reproducibility. Data post-processing was also focused on integrating user comfort perception with image data, and it revealed anomaly detections represented by numerical acceleration spikes. Additionally, data positioning was successfully implemented. Finally, awz measurements with GPS coordinates and images were incorporated into a Geographic Information System (GIS) to develop a database that supports the efficient and comprehensive management of surface conditions. The proposed system can be considered as a valuable tool to assess the pavement conditions of cycle paths in order to implement preventive maintenance strategies within budget constraints. Full article
Show Figures

Figure 1

Figure 1
<p>Flowchart of proposed methodology.</p>
Full article ">Figure 2
<p>The proposed cycle path monitoring system.</p>
Full article ">Figure 3
<p>The placement of the core hardware setup on the bicycle’s top tube.</p>
Full article ">Figure 4
<p>The imaging system. (<b>a</b>) The placement of the camera module on the handlebars; (<b>b</b>) a close-up view of the custom, 3D-printed mount designed to attach the module.</p>
Full article ">Figure 5
<p>The IMU was fixed inside the bicycle’s saddle.</p>
Full article ">Figure 6
<p>The GPS system. (<b>a</b>) The installation of the u-blox NEO-6M GPS module beneath the bicycle saddle; (<b>b</b>) a close-up view of the GPS module mounted on the custom, 3D-printed bracket.</p>
Full article ">Figure 7
<p>(<b>a</b>) The placement of the magnets on the spokes of the bicycle’s rear wheel for the Hall Effect sensor; (<b>b</b>) a close-up view of the Hall Effect sensor module mounted near the rear wheel, aligned to detect the passing magnets for accurate distance measurement.</p>
Full article ">Figure 8
<p>The 8 km of Rome’s cycle path network examined in the field test. The selected branches are identified according to <a href="#infrastructures-10-00050-t005" class="html-table">Table 5</a>.</p>
Full article ">Figure 9
<p>Correction of GPS trajectory discrepancies.</p>
Full article ">Figure 10
<p>Validation of proposed cycle path monitoring. The different colors in the graph area correspond to comfort classes.</p>
Full article ">Figure 11
<p>Schematic representation of camera’s field of view.</p>
Full article ">Figure 12
<p><math display="inline"><semantics> <mrow> <msub> <mrow> <mi>a</mi> </mrow> <mrow> <mi>w</mi> <mi>z</mi> </mrow> </msub> </mrow> </semantics></math> spike of 1.59 m/s<sup>2</sup> at 25th second on Branch 4.</p>
Full article ">Figure 13
<p>The drainage grate identified as the cause of the <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>a</mi> </mrow> <mrow> <mi>w</mi> <mi>z</mi> </mrow> </msub> </mrow> </semantics></math> spike in <a href="#infrastructures-10-00050-f010" class="html-fig">Figure 10</a>.</p>
Full article ">Figure 14
<p>Example of integrated data in QGIS, displaying GPS coordinates, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>a</mi> </mrow> <mrow> <mi>w</mi> <mi>z</mi> </mrow> </msub> </mrow> </semantics></math> values, and corresponding video frames for sample unit of investigated cycle path. The red dot indicates which sample unit is under investigation.</p>
Full article ">Figure 15
<p>Visualization of <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>a</mi> </mrow> <mrow> <mi>w</mi> <mi>z</mi> </mrow> </msub> </mrow> </semantics></math> values along Lungotevere cycle path.</p>
Full article ">
18 pages, 7121 KiB  
Article
Single-Model Self-Recovering Fringe Projection Profilometry Absolute Phase Recovery Method Based on Deep Learning
by Xu Li, Yihao Shen, Qifu Meng, Mingyi Xing, Qiushuang Zhang and Hualin Yang
Sensors 2025, 25(5), 1532; https://doi.org/10.3390/s25051532 - 1 Mar 2025
Viewed by 191
Abstract
A drawback of fringe projection profilometry (FPP) is that it is still a challenge to perform efficient and accurate high-resolution absolute phase recovery with only a single measurement. This paper proposes a single-model self-recovering fringe projection absolute phase recovery method based on deep [...] Read more.
A drawback of fringe projection profilometry (FPP) is that it is still a challenge to perform efficient and accurate high-resolution absolute phase recovery with only a single measurement. This paper proposes a single-model self-recovering fringe projection absolute phase recovery method based on deep learning. The built Fringe Prediction Self-Recovering network converts a single fringe image acquired by a camera into four single mode self-recovering fringe images. A self-recovering algorithm is adopted to obtain wrapped phases and fringe grades, realizing high-resolution absolute phase recovery from only a single shot. Low-cost and efficient dataset preparation is realized by the constructed virtual measurement system. The fringe prediction network showed good robustness and generalization ability in experiments with multiple scenarios using different lighting conditions in both virtual and physical measurement systems. The absolute phase recovered MAE in the real physical measurement system was controlled to be 0.015 rad, and the reconstructed point cloud fitting RMSE was 0.02 mm. It was experimentally verified that the proposed method can achieve efficient and accurate absolute phase recovery under complex ambient lighting conditions. Compared with the existing methods, the method in this paper does not need the assistance of additional modes to process the high-resolution fringe images directly. Combining the deep learning technique with the self-recovering algorithm simplified the complex process of phase retrieval and phase unwrapping, and the proposed method is simpler and more efficient, which provides a reference for the fast, lightweight, and online detection of FPP. Full article
(This article belongs to the Section Optical Sensors)
Show Figures

Figure 1

Figure 1
<p>Flow diagram for the single-model self-recovering phase fringe projection technique using FPSR-Net for absolute phase recovery (Colored lines in deep learning part represent the trend of light intensity distribution in fringe images).</p>
Full article ">Figure 2
<p>Fringe prediction self-recovering phase network. (<b>a</b>) Overall network architecture; (<b>b</b>) fringe prediction dual attention block; (<b>c</b>) position attention model; (<b>d</b>) channel attention model; (<b>e</b>) single transformer layer.</p>
Full article ">Figure 3
<p>Schematic diagram of measurement system. (<b>a</b>) Physical measurement system (only one camera was used in this paper); (<b>b</b>) simulated measurement system.</p>
Full article ">Figure 4
<p>Example sets of three scenarios in the dataset: (<b>a</b>,<b>f</b>,<b>k</b>) ideal light intensity; (<b>b</b>,<b>g</b>,<b>l</b>) minimum light intensity; (<b>c</b>,<b>h</b>,<b>m</b>) maximum light intensity; (<b>d</b>,<b>i</b>,<b>n</b>) ground truth; (<b>e</b>,<b>j</b>,<b>o</b>) absolute phase of ground truth recovery.</p>
Full article ">Figure 5
<p>Image path segmentation, synthesis, and comparison. (<b>a</b>) Schematic of image segmentation and synthesis (The red solid line is the start position of each patch block segmentation, and the red dashed line is the end position. The yellow and blue areas represent the R1C2 patch and the R2C3 patch, respectively); (<b>b</b>) comparison of predicted results combined with ground truth.</p>
Full article ">Figure 6
<p>The loss curve of the training stage.</p>
Full article ">Figure 7
<p>Absolute phase measurement error results for three scenarios (The red box represents a partial zoomed-in view).</p>
Full article ">Figure 8
<p>Absolute phase measurement error results for different modules under dim ambient light interference.</p>
Full article ">Figure 9
<p>Absolute phase measurement error results for different modules under exposure ambient light interference.</p>
Full article ">Figure 10
<p>Measurement results under actual ambient light interference. (<b>a</b>) Exposure scene, (<b>b</b>) dim scene, (<b>c</b>) reflection scene. (The red color indicates the location of the phase defect).</p>
Full article ">Figure 11
<p>Measurement results of the dynamic scene. (<b>a</b>–<b>d</b>) Images and measurement results of proposed FPSR-Net at different times.</p>
Full article ">
21 pages, 21131 KiB  
Article
Measurement of Human Body Segment Properties Using Low-Cost RGB-D Cameras
by Cristina Nuzzi, Marco Ghidelli, Alessandro Luchetti, Matteo Zanetti, Francesco Crenna and Matteo Lancini
Sensors 2025, 25(5), 1515; https://doi.org/10.3390/s25051515 - 28 Feb 2025
Viewed by 208
Abstract
An open question for the biomechanical research community is accurate estimation of the volume and mass of each body segment of the human body, especially when indirect measurements are based on biomechanical modeling. Traditional methods involve the adoption of anthropometric tables, which describe [...] Read more.
An open question for the biomechanical research community is accurate estimation of the volume and mass of each body segment of the human body, especially when indirect measurements are based on biomechanical modeling. Traditional methods involve the adoption of anthropometric tables, which describe only the average human shape, or manual measurements, which are time-consuming and depend on the operator. We propose a novel method based on the acquisition of a 3D scan of a subject’s body, which is obtained using a consumer-end RGB-D camera. The body segments’ separation is obtained by combining the body skeleton estimation of BlazePose with a biomechanical-coherent skeletal model, which is defined according to the literature. The volume of each body segment is computed using a 3D Monte Carlo procedure. Results were compared with manual measurement by experts, anthropometric tables, and a model leveraging truncated cone approximations, showing good adherence to reference data with minimal differences (ranging from +0.5 to 1.0 dm3 for the upper limbs, 0.1 to 4.2 dm3 for the thighs, and 0.4 to 2.3 dm3 for the shanks). In addition, we propose a novel indicator based on the computation of equivalent diameters for each body segment, highlighting the importance of gender-specific biomechanical models to account for the chest and pelvis areas of female subjects. Full article
(This article belongs to the Special Issue 3D Reconstruction with RGB-D Cameras and Multi-sensors)
Show Figures

Figure 1

Figure 1
<p>Scheme of the complete processing procedure with example results taken from one subject’s data.</p>
Full article ">Figure 2
<p>(<b>a</b>) Scheme of the KPs considered for this work. Red KPs: computed by BlazePose; yellow KPs: calculated as midpoints of shoulders and hips segments; and black KPs: obtained by applying the biomechanical model to the original BlazePose KPs. (<b>b</b>) An example of vectors that were computed by the biomechanical model.</p>
Full article ">Figure 3
<p>Examples of the filtering process. (<b>a</b>) Original point cloud, <math display="inline"><semantics> <mrow> <mi>P</mi> <msub> <mi>C</mi> <mrow> <mi>r</mi> <mi>a</mi> <mi>w</mi> </mrow> </msub> </mrow> </semantics></math>. (<b>b</b>) Result of the coarse filtering process, <math display="inline"><semantics> <mrow> <mi>P</mi> <msub> <mi>C</mi> <mrow> <mi>c</mi> <mi>o</mi> <mi>a</mi> <mi>r</mi> <mi>s</mi> <mi>e</mi> <mo>,</mo> <mn>0</mn> </mrow> </msub> </mrow> </semantics></math>. (<b>c</b>) Result of the fine filtering process, <math display="inline"><semantics> <mrow> <mi>P</mi> <msub> <mi>C</mi> <mrow> <mi>f</mi> <mi>i</mi> <mi>n</mi> <mi>e</mi> </mrow> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 4
<p>Examples of the orientation correction of the fine filtering process. (<b>a</b>) Alignment of the bed’s normal to the Z reference axis. (<b>b</b>) Alignment of the bed’s Monte Carlo approximation principal components to the X and Y reference axes.</p>
Full article ">Figure 5
<p>Example of some EPs drawn on a subject’s point cloud projected onto a 2D XY plane. Each EP is shown in black, and the points belonging to them are highlighted in different colors. Vectors used to create the remaining EPs are shown in red.</p>
Full article ">Figure 6
<p>Examples of the body segment separation procedure. (<b>a</b>) <math display="inline"><semantics> <mrow> <mi>P</mi> <msub> <mi>C</mi> <mrow> <mi>f</mi> <mi>i</mi> <mi>n</mi> <mi>e</mi> </mrow> </msub> </mrow> </semantics></math> of a subject on top, of which the 19 KPs are drawn in green. (<b>b</b>) Separation of the BSs of the subject highlighted in different colors. (<b>c</b>) Monte Carlo approximation of the BSs, which is necessary to estimate their volume.</p>
Full article ">Figure 7
<p>(<b>a</b>) Point cloud filling example of the shoulder’s BS. Red: original points; green: artificial points created to fill vertical gaps around pointy areas; and cyan: artificial points created to fill the bottom part of the BS. (<b>b</b>) Alpha shape computed on the shoulder’s BS. (<b>c</b>) Monte Carlo points of the shoulder’s BS belonging to its alpha shape.</p>
Full article ">Figure 8
<p>Images of the subjects with section lines that were directly drawn by an expert on the skin using a pen for the female subject and a tape for the male subject (which is highlighted in the picture). Vectors computed from the KPs were superimposed on the image to highlight any discrepancy. (<b>a</b>) Male subject, (<b>b</b>) female subject.</p>
Full article ">Figure 9
<p>Boxplot of the resulting <math display="inline"><semantics> <mrow> <mo>Δ</mo> <mi>V</mi> </mrow> </semantics></math> values. Comparison was only made for the limbs. Blue boxes refer to male data and pink boxes to female data. (<b>a</b>) Difference between the scanner volume and the one obtained from the literature reference, <math display="inline"><semantics> <mrow> <mo>Δ</mo> <msub> <mi>V</mi> <mrow> <mi>s</mi> <mi>c</mi> <mi>a</mi> <mi>n</mi> <mi>n</mi> <mi>e</mi> <mi>r</mi> <mo>−</mo> <mi>r</mi> <mi>e</mi> <mi>f</mi> <mo>,</mo> <mi>B</mi> <mi>S</mi> </mrow> </msub> </mrow> </semantics></math>. (<b>b</b>) Difference between the scanner volume and the one obtained by approximating the BS to a truncated cone using the subject’s manual measurements, <math display="inline"><semantics> <mrow> <mo>Δ</mo> <msub> <mi>V</mi> <mrow> <mi>s</mi> <mi>c</mi> <mi>a</mi> <mi>n</mi> <mi>n</mi> <mi>e</mi> <mi>r</mi> <mo>−</mo> <mi>c</mi> <mi>o</mi> <mi>n</mi> <mi>e</mi> <mo>,</mo> <mi>B</mi> <mi>S</mi> </mrow> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 10
<p>Boxplot of the resulting <math display="inline"><semantics> <mrow> <mo>Δ</mo> <mi>L</mi> </mrow> </semantics></math> values, which were normalized over the subject’s height (SH). Blue boxes refer to male data and pink boxes to female data. Outliers are depicted in black with a plus symbol. (<b>a</b>) Difference between the BS lengths obtained from our biomechanical model and the BS lengths obtained from the literature reference, <math display="inline"><semantics> <mrow> <mo>Δ</mo> <msub> <mi>L</mi> <mrow> <mi>s</mi> <mi>c</mi> <mi>a</mi> <mi>n</mi> <mi>n</mi> <mi>e</mi> <mi>r</mi> <mo>−</mo> <mi>r</mi> <mi>e</mi> <mi>f</mi> <mo>,</mo> <mi>B</mi> <mi>S</mi> </mrow> </msub> </mrow> </semantics></math>. (<b>b</b>) Difference between the BS lengths obtained from our biomechanical model and the BS lengths obtained from manual measurements, <math display="inline"><semantics> <mrow> <mo>Δ</mo> <msub> <mi>L</mi> <mrow> <mi>s</mi> <mi>c</mi> <mi>a</mi> <mi>n</mi> <mi>e</mi> <mi>r</mi> <mo>−</mo> <mi>c</mi> <mi>o</mi> <mi>n</mi> <mi>e</mi> <mo>,</mo> <mi>B</mi> <mi>S</mi> </mrow> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 11
<p>(<b>a</b>) Boxplot of the <math display="inline"><semantics> <mrow> <mo>Δ</mo> <msub> <mi>D</mi> <mrow> <mi>B</mi> <mi>S</mi> </mrow> </msub> </mrow> </semantics></math> computed as the difference between <math display="inline"><semantics> <msub> <mover accent="true"> <mi>D</mi> <mo stretchy="false">^</mo> </mover> <mrow> <mi>s</mi> <mi>c</mi> <mi>a</mi> <mi>n</mi> <mo>,</mo> <mi>B</mi> <mi>S</mi> </mrow> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mover accent="true"> <mi>D</mi> <mo stretchy="false">^</mo> </mover> <mrow> <mi>r</mi> <mi>e</mi> <mi>f</mi> <mo>,</mo> <mi>B</mi> <mi>S</mi> </mrow> </msub> </semantics></math>. The difference was normalized over the subjects’ height (SH). (<b>b</b>) Difference of the BSs mass resulting from our scanner-based method and the reference tables. The data were normalized over the subject’s body weight (BW).</p>
Full article ">Figure 12
<p>(<b>a</b>) Histogram distribution of the normalized diameters for both the rescaled table values (red) and manual measurements (blue) for the leg BS. (<b>b</b>) Bland–Altman plot of the normalized diameters of the thigh BS.</p>
Full article ">
21 pages, 27582 KiB  
Article
Multi-Level Spectral Attention Network for Hyperspectral BRDF Reconstruction from Multi-Angle Multi-Spectral Images
by Liyao Song and Haiwei Li
Remote Sens. 2025, 17(5), 863; https://doi.org/10.3390/rs17050863 - 28 Feb 2025
Viewed by 185
Abstract
With the rapid development of hyperspectral applications using unmanned aerial vehicles (UAVs), the traditional assumption that ground objects exhibit Lambertian reflectance is no longer sufficient to meet the high-precision requirements for quantitative inversion and airborne hyperspectral data applications. Therefore, it is necessary to [...] Read more.
With the rapid development of hyperspectral applications using unmanned aerial vehicles (UAVs), the traditional assumption that ground objects exhibit Lambertian reflectance is no longer sufficient to meet the high-precision requirements for quantitative inversion and airborne hyperspectral data applications. Therefore, it is necessary to establish a hyperspectral bidirectional reflectance distribution function (BRDF) model suitable for the area of imaging. However, obtaining multi-angle information from UAV push-broom hyperspectral data is difficult. Achieving uniform push-broom imaging and flexibly acquiring multi-angle data is challenging due to spatial distortions, particularly under heightened roll or pitch angles, and the need for multiple flights; this extends acquisition time and exacerbates uneven illumination, introducing errors in BRDF model construction. To address these issues, we propose leveraging the advantages of multi-spectral cameras, such as their compact size, lightweight design, and high signal-to-noise ratio (SNR) to reconstruct hyperspectral multi-angle data. This approach enhances spectral resolution and the number of bands while mitigating spatial distortions and effectively captures the multi-angle characteristics of ground objects. In this study, we collected UAV hyperspectral multi-angle data, corresponding illumination information, and atmospheric parameter data, which can solve the problem of existing BRDF modeling not considering outdoor ambient illumination changes, as this limits modeling accuracy. Based on this dataset, we propose an improved Walthall model, considering illumination variation. Then, the radiance consistency of BRDF multi-angle data is effectively optimized, the error caused by illumination variation in BRDF modeling is reduced, and the accuracy of BRDF modeling is improved. In addition, we adopted Transformer for spectral reconstruction, increased the number of bands on the basis of spectral dimension enhancement, and conducted BRDF modeling based on the spectral reconstruction results. For the multi-level Transformer spectral dimension enhancement algorithm, we added spectral response loss constraints to improve BRDF accuracy. In order to evaluate BRDF modeling and quantitative application potential from the reconstruction results, we conducted comparison and ablation experiments. Finally, we solved the problem of difficulty in obtaining multi-angle information due to the limitation of hyperspectral imaging equipment, and we provide a new solution for obtaining multi-angle features of objects with higher spectral resolution using low-cost imaging equipment. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

Figure 1
<p>Multi-level BRDF spectral reconstruction network. (<b>a</b>) Single-level spectral transformer module SST; (<b>b</b>) Multi-level spectral reconstruction network.</p>
Full article ">Figure 2
<p>The structure of each component in the SST module. (<b>a</b>) Spectral Multi-Head Attention Module S-MSA; (<b>b</b>) Dual RsFFN; (<b>c</b>) Spectral Attention Module SAB.</p>
Full article ">Figure 3
<p>UAV nested multi-rectangular flight routes.</p>
Full article ">Figure 4
<p>Changes in aerosol and water vapor content on the day of the experiment (<b>left column</b>: the first day; <b>right column</b>: the second day).</p>
Full article ">Figure 5
<p>Schematic diagram of observation angles at the moment of UAV imaging.</p>
Full article ">Figure 6
<p>Processing flow of multi-angle data.</p>
Full article ">Figure 7
<p>Comparison of true color results of BRDF data reconstructed using different methods at different observation zenith angles.</p>
Full article ">Figure 8
<p>Comparison of mean spectral curves and error curves reconstructed by different BRDF methods at different angles.</p>
Full article ">Figure 9
<p>Error heat map comparison of the reconstruction results of different reconstruction methods in the 5-<math display="inline"><semantics> <mrow> <mi>t</mi> <mi>h</mi> </mrow> </semantics></math> and 15-<math display="inline"><semantics> <mrow> <mi>t</mi> <mi>h</mi> </mrow> </semantics></math> bands.</p>
Full article ">Figure 10
<p>Comparison of Walthall model data distribution with/without considering the illumination variation.</p>
Full article ">Figure 11
<p>Hyperspectral BRDF modeling with illumination correction Walthall model.</p>
Full article ">Figure 12
<p>Multi-angle spectral reconstruction BRDF modeling considering the illumination-corrected Walthall model.</p>
Full article ">Figure 13
<p>Error analysis comparison between spectral reconstruction BRDF model and hyperspectral BRDF model.</p>
Full article ">
19 pages, 10608 KiB  
Article
Urban Waterlogging Monitoring and Recognition in Low-Light Scenarios Using Surveillance Videos and Deep Learning
by Jian Zhao, Xing Wang, Cuiyan Zhang, Jing Hu, Jiaquan Wan, Lu Cheng, Shuaiyi Shi and Xinyu Zhu
Water 2025, 17(5), 707; https://doi.org/10.3390/w17050707 - 28 Feb 2025
Viewed by 239
Abstract
With the intensification of global climate change, extreme precipitation events are occurring more frequently, making the monitoring and management of urban flooding a critical global issue. Urban surveillance camera sensor networks, characterized by their large-scale deployment, rapid data transmission, and low cost, have [...] Read more.
With the intensification of global climate change, extreme precipitation events are occurring more frequently, making the monitoring and management of urban flooding a critical global issue. Urban surveillance camera sensor networks, characterized by their large-scale deployment, rapid data transmission, and low cost, have emerged as a key complement to traditional remote sensing techniques. These networks offer new opportunities for high-spatiotemporal-resolution urban flood monitoring, enabling real-time, localized observations that satellite and aerial systems may not capture. However, in low-light environments—such as during nighttime or heavy rainfall—the image features of flooded areas become more complex and variable, posing significant challenges for accurate flood detection and timely warnings. To address these challenges, this study develops an imaging model tailored to flooded areas under low-light conditions and proposes an invariant feature extraction model for flooding areas within surveillance videos. By using extracted image features (i.e., brightness and invariant features of flooded areas) as inputs, a deep learning-based flood segmentation model is built on the U-Net architecture. A new low-light surveillance flood image dataset, named UWs, is constructed for training and testing the model. The experimental results demonstrate the efficacy of the proposed method, achieving an mRecall of 0.88, an mF1_score of 0.91, and an mIoU score of 0.85. These results significantly outperform the comparison algorithms, including LRASPP, DeepLabv3+ with MobileNet and ResNet backbones, and the classic DeepLabv3+, with improvements of 4.9%, 3.0%, and 4.4% in mRecall, mF1_score, and mIoU, respectively, compared to Res-UNet. Additionally, the method maintains its strong performance in real-world tests, and it is also effective for daytime flood monitoring, showcasing its robustness for all-weather applications. The findings of this study provide solid support for the development of an all-weather urban surveillance camera flood monitoring network, with significant practical value for enhancing urban emergency management and disaster reduction efforts. Full article
(This article belongs to the Section Urban Water Management)
Show Figures

Figure 1

Figure 1
<p>The infrared images of the surveillance camera. (<b>a</b>–<b>d</b>) show the changes in the water features before and after the car drives by. (<b>e</b>–<b>h</b>) show the changes in water features at different surveillance camera positions and attitudes. The flooding area is labeled in red.</p>
Full article ">Figure 2
<p>False color images of the surveillance camera. (<b>a</b>–<b>d</b>) show the changes in the water features before and after a car drives by. (<b>e</b>–<b>h</b>) show the variation in water features at different surveillance camera positions and poses. The flooding area is labeled in red.</p>
Full article ">Figure 3
<p>Architecture of Aunet. The Dis<sub>LA</sub> module realizes the separation of the invariant features of the flooding area of the low-light scene, and the U-Net [<a href="#B43-water-17-00707" class="html-bibr">43</a>] network realizes the segmentation of the flooding region of the low-light scene. The white areas represent flood.</p>
Full article ">Figure 4
<p>Dis<sub>LA</sub> composition of the module, where 7 × 7 @128 denotes that there are 128 sets of convolutional kernels of the size 7 × 7. * [N, H, W, C] denotes the input dimension of the module, where N is the Batchsize, H is the height, W is the width, and C is the number of channels.</p>
Full article ">Figure 5
<p>The architecture of the Swin Transformer block [<a href="#B47-water-17-00707" class="html-bibr">47</a>].</p>
Full article ">Figure 6
<p>Some surveillance scenarios within the constructed dataset.</p>
Full article ">Figure 7
<p>Segmentation effect of floods for each model in the black-and-white image. The red areas represent floods. The green-framed areas are of particular interest.</p>
Full article ">Figure 8
<p>Segmentation effect of flooding areas for each model in the false color image. The red areas represent floods. The green, yellow, and blue-framed areas are of particular interest.</p>
Full article ">Figure 9
<p>Effect of Aunet and comparison models on negative samples (water-free regions at night). The red areas represent floods.</p>
Full article ">Figure 10
<p>Aunet and comparison modeling of flooding area segmentation effect in daytime flood images. The red areas represent floods.</p>
Full article ">
22 pages, 16205 KiB  
Article
Hyper Spectral Camera ANalyzer (HyperSCAN)
by Wen-Qian Chang, Hsun-Ya Hou, Pei-Yuan Li, Michael W. Shen, Cheng-Ling Kuo, Tang-Huang Lin, Loren C. Chang, Chi-Kuang Chao and Jann-Yenq Liu
Remote Sens. 2025, 17(5), 842; https://doi.org/10.3390/rs17050842 - 27 Feb 2025
Viewed by 130
Abstract
HyperSCAN (Hyper Spectral Camera ANalyzer) is a hyperspectral imager which monitors the Earth’s environment and also an educational platform to integrate college students’ ideas and skills in optical design and data processing. The advantages of HyperSCAN are that it is designed for modular [...] Read more.
HyperSCAN (Hyper Spectral Camera ANalyzer) is a hyperspectral imager which monitors the Earth’s environment and also an educational platform to integrate college students’ ideas and skills in optical design and data processing. The advantages of HyperSCAN are that it is designed for modular design, is compact and lightweight, and low-cost using commercial off-the-shelf (COTS) optical components. The modular design allows for flexible and rapid development, as well as validation within college lab environments. To optimize space utilization and reduce the optical path, HyperSCAN’s optical system incorporates a folding mirror, making it ideal for the constrained environment of a CubeSat. The use of COTS components significantly lowers pre-development costs and minimizes associated risks. The compact size and cost-effectiveness of CubeSats, combined with the advanced capabilities of hyperspectral imagers, make them a powerful tool for a broad range of applications, such as environmental monitoring of Earth, disaster management, mineral and resource exploration, atmospheric and climate studies, and coastal and marine research. We conducted a spatial-resolution-boost experiment using HyperSCAN data and various hyperspectral datasets including Urban, Pavia University, Pavia Centre, Botswana, and Indian Pines. After testing various data-fusion deep learning models, the best image quality of these methods is a two-branches convolutional neural network (TBCNN), where TBCNN retrieves spatial and spectral features in parallel and reconstructs the higher-spatial-resolution data. With the aid of higher-spatial-resolution multispectral data, we can boost the spatial resolution of HyperSCAN data. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) The mechanism design, and (<b>b</b>) the flight model of HyperSCAN.</p>
Full article ">Figure 2
<p>Rendering of the structural design of the SCION-X 12U CubeSat.</p>
Full article ">Figure 3
<p>The HyperSCAN is composed of the following optical components: (1) the front camera lens assembly, (2) the slit in the focal plane, (3) the folding mirror, (4) diffractive element (concave grating), and (5) the imaging sensor.</p>
Full article ">Figure 4
<p>(<b>a</b>) The slice scans of the Earth surface with a swath width of 80.3 km (red double-arrow line), and its along-track length of 33.3 km, which is illustrated by a red thick arrow. A black thick arrow indicates a flying direction of HyperSCAN at the CubeSat orbit height H, (<b>b</b>) 2D sliced image with spatial and spectral information of Earth’s surface, and (<b>c</b>) 3D hyperspectral data cube composed by successive line scans of ground view.</p>
Full article ">Figure 5
<p>(<b>a</b>) An RGB image reconstructed from three selected bands of a hyperspectral image captured during the HyperSCAN experiment, with regions of interest highlighted with green, blue, red, and cyan filled circles. (<b>b</b>) The spectral radiance curves of different materials, including grass, sky, buildings, and glass, over the wavelength range of 450 nm to 850 nm.</p>
Full article ">Figure 6
<p>The mechanical design of the camera at (<b>a</b>) the transparent view and (<b>b</b>) the 3D view. At the transparent view, a lens system of camera has 4 elements (labeled by G1, G2, G3, and G4).</p>
Full article ">Figure 7
<p>The functional block diagram of subsystems in the ICDH of HyperSCAN.</p>
Full article ">Figure 8
<p>The data interfaces between subsystems in the ICDH of HyperSCAN.</p>
Full article ">Figure 9
<p>The power interfaces between subsystems in the ICDH of HyperSCAN.</p>
Full article ">Figure 10
<p>Two frameworks and approaches using deep learning data reconstruction models: (<b>a</b>) feature-input fusion and (<b>b</b>) input-level fusion.</p>
Full article ">Figure 11
<p>Spatial-resolution-boost experiments from low-spectral-resolution MSI and low-spatial-resolution HSI data using TBCNN, SSRNET, SpatCNN, SpecCNN, ConSSFCNN, SSFCNN, ResTFNet, TFNet, and MSDCNN where TBCNN demonstrates the minimum differences at two edge wavelength (450 and 850 nm).</p>
Full article ">Figure 12
<p>Spatial-resolution-boost experiments using low-spectral-resolution multispectral and hyperspectral data.</p>
Full article ">Figure 13
<p>Reconstructed data at various wavelength from TBCNN in the third row in comparison with low-spatial-resolution HSI in the first row and original high-spatial-resolution HSI in the second row. The TBCNN greatly improves the spatial resolution of low-spatial-resolution HSI along with low-spectral-resolution MSI.</p>
Full article ">Figure 14
<p>Spatial-resolution-boost experiments on the Urban dataset: RGB pseudocolor visualization of TBCNN, SSRNET, SpatCNN, SpecCNN, ConSSFCNN, SSFCNN, ResTFNet, TFNet, and MSDCNN.</p>
Full article ">Figure 15
<p>Spatial-resolution-boost experiments on the Pavia University dataset: RGB pseudo-color visualization of TBCNN, SSRNET, SpatCNN, SpecCNN, ConSSFCNN, SSFCNN, ResTFNet, TFNet, and MSDCNN.</p>
Full article ">Figure 16
<p>Spatial-resolution-boost experiments on the Pavia Centre dataset: RGB pseudocolor visualization of TBCNN, SSRNET, SpatCNN, SpecCNN, ConSSFCNN, SSFCNN, ResTFNet, TFNet, and MSDCNN.</p>
Full article ">Figure 17
<p>Spatial-resolution-boost experiments on the Botwana dataset: RGB pseudocolor visualization of TBCNN, SSRNET, SpatCNN, SpecCNN, ConSSFCNN, SSFCNN, ResTFNet, TFNet, and MSDCNN.</p>
Full article ">Figure 18
<p>Spatial-resolution-boost experiments on the Indian Pines dataset: RGB pseudocolor visualization of TBCNN, SSRNET, SpatCNN, SpecCNN, ConSSFCNN, SSFCNN, ResTFNet, TFNet, and MSDCNN.</p>
Full article ">
7 pages, 4095 KiB  
Brief Report
Hive Insulation Increases Foraging Activities of Bumble Bees (Bombus impatiens) in a Wild Blueberry Field in Quebec, Canada
by Maxime C. Paré, Nasimeh Mortazavi, Jean-Denis Brassard, Thierry Chouffot, Julie Douillard and G. Christopher Cutler
Agronomy 2025, 15(3), 562; https://doi.org/10.3390/agronomy15030562 - 25 Feb 2025
Viewed by 300
Abstract
Common eastern bumble bees (Bombus impatiens Cresson) play an essential role in pollinating lowbush blueberries (LB) in northern Quebec, but their costs and the suboptimal weather conditions during pollination highlight the need to find appropriate hive management strategies. A study was conducted [...] Read more.
Common eastern bumble bees (Bombus impatiens Cresson) play an essential role in pollinating lowbush blueberries (LB) in northern Quebec, but their costs and the suboptimal weather conditions during pollination highlight the need to find appropriate hive management strategies. A study was conducted in a LB field in Saguenay (Québec, Canada) focusing on the effects of hive insulation (I+ and I−), heating (H+ and H−), and placement in a single-row tree line windbreak. High-definition time-lapse cameras monitored hive activities and bumble bee foraging behaviors. We found that the conventional management of placing hives in full sun without insulation (I−) resulted in the lowest levels of bumble bee foraging activity and overall hive traffic. Placing bumble bee hives against a windbreak resulted in the highest numbers of bees entering hives with pollen (+156%), leaving hives (+69%), and overall hive traffic (+76%). Insulating hives with extruded polystyrene foam gave intermediate results, with a 105% increase in foraging activity compared to the conventional management method (I−H−). Interestingly, placing hives on seedling mats to maintain colony temperatures above 15 °C (H+) tended to decrease foraging activity and overall hive traffic. Our results show that strategic placement of bumble bee hives against windbreaks can significantly increase the activity of Bombus workers from those hives and can be used as a simple, low-cost, and efficient bumble bee hive management method by LB growers. Full article
(This article belongs to the Section Horticultural and Floricultural Crops)
Show Figures

Figure 1

Figure 1
<p>The four management treatments of bumble bee hives (Quad<sup>®</sup>, Koppert, Scarborough, ON, Canada). (<b>A</b>) Insulated and heated hives (I+H+). (<b>B</b>) Insulated and unheated hives (I+H−). (<b>C</b>) Hives placed against a single-row windbreak. (<b>D</b>) Uninsulated and unheated hives (I−H−). Cameras were attached to a wood stick and placed at 1.2 m from the hives to monitor bumblebee activity from the two facing hives.</p>
Full article ">Figure 2
<p>Effects of hive management treatments (<b>A</b>–<b>C</b>), outside temperature (°C) ranges (<b>D</b>–<b>F</b>), and weather conditions (<b>G</b>–<b>I</b>) on bumble bees entering in the hive with pollen on their hind legs (<b>A</b>,<b>D</b>,<b>G</b>), leaving out the hive (<b>B</b>,<b>E</b>,<b>H</b>), and hive overall traffic (<b>C</b>,<b>F</b>,<b>I</b>). Results were collected between 9 a.m. and 5 p.m. from 21 May to 8 June 2021. Columns sharing letters are not significantly different according to the non-parametric Kruskal–Wallis test (α = 0.05). Error bars represent standard error from the mean. Insulating (I+), not insulating (I−), heating (H+), not heating (H−), and against the windbreak (windbreak).</p>
Full article ">
27 pages, 6382 KiB  
Article
Utilizing IoT Sensors and Spatial Data Mining for Analysis of Urban Space Actors’ Behavior in University Campus Space Design
by Krzysztof Koszewski, Robert Olszewski, Piotr Pałka, Renata Walczak, Przemysław Korpas, Karolina Dąbrowska-Żółtak, Michał Wyszomirski, Olga Czeranowska-Panufnik, Andrzej Manujło, Urszula Szczepankowska-Bednarek, Joanna Kuźmicz-Kubiś, Anna Szalwa, Krzysztof Ejsmont and Paweł Czernic
Sensors 2025, 25(5), 1393; https://doi.org/10.3390/s25051393 - 25 Feb 2025
Viewed by 459
Abstract
This paper discusses the use of IoT sensor networks and spatial data mining methods to support the design process in the revitalization of the university campus of the Warsaw University of Technology (WUT) in the spirit of universal design. The aim of the [...] Read more.
This paper discusses the use of IoT sensor networks and spatial data mining methods to support the design process in the revitalization of the university campus of the Warsaw University of Technology (WUT) in the spirit of universal design. The aim of the research was to develop a methodology for the use of IoT and edge computing for the acquisition of spatial knowledge based on spatial big data, as well as for the development of an open (geo)information society that shares the responsibility for the process of shaping the spaces of smart cities. The purpose of the article is to verify the hypothesis on whether it is possible to obtain spatial–temporal quantitative data that are useful in the process of designing the space of a university campus using low-cost Internet of Things sensors, i.e., already existing networks of CCTV cameras supported by simple installed beam-crossing sensors. The methodological approach proposed in the article combines two main areas—the use of IT technologies (IoT, big data, spatial data mining) and data-driven design based on analysis of urban space actors’ behavior for participatory revitalization of a university campus. The research method applied involves placing a network of locally communicating heterogeneous IoT sensors in the space of a campus. These sensors collect data on the behavior of urban space actors: people and vehicles. The data collected and the knowledge gained from its analysis are used to discuss the shape of the campus space. The testbed of the developed methodology was the central campus of the WUT (Warsaw University of Technology), which made it possible to analyze the time-varying use of the selected campus spaces and to identify the premises for the revitalization project in accordance with contemporary trends in the design of the space of HEIs (higher education institutions), as well as the needs of the academic community and the residents of the capital. The results are used not only to optimize the process of redesigning the WUT campus, but also to support the process of discussion and activation of the community in the development of deliberative democracy and participatory shaping of space in general. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

Figure 1
<p>An urban map illustrating the campus location within the city, accompanied by a campus map highlighting the specific area where measurements and analyses were conducted.</p>
Full article ">Figure 2
<p>A simplified map of the Central Campus of the WUT, with pedestrian and bicycle entrances and vehicle entrances marked.</p>
Full article ">Figure 3
<p>Acquisition process of spatial–temporal knowledge seen from two perspectives: (<b>a</b>) our IoT sensor network’s architecture and (<b>b</b>) a data processing scheme.</p>
Full article ">Figure 4
<p>Location of cameras and sensors on a map of the WUT Main Campus.</p>
Full article ">Figure 5
<p>The view from each of the analyzed surveillance cameras.</p>
Full article ">Figure 6
<p>An example of a single frame from the Golski camera with detected objects (pedestrians and cars) being indicated by rectangles. Small circles indicate the trajectories of their movements.</p>
Full article ">Figure 7
<p>An example of trajectory visualization in the Kepler.gl application: pedestrians (<b>a</b>) and cars (<b>b</b>).</p>
Full article ">Figure 8
<p>An example of data structure (presented in a JSON-like format) describing the passage of a person through a square with X, Y coordinates: 7500559, 5787471.</p>
Full article ">Figure 9
<p>Explanation of the technique to automatically remove the time shift between beam-crossing and CCTV sensors: positions of physical and virtual beam-crossing sensors (<b>a</b>), initial signal vector created from detected events (<b>b</b>), both signals after low-pass filtering (<b>c</b>), result of cross-correlation between those signals (<b>d</b>), events from both sensors after time alignment (<b>e</b>) (“Golski” camera, 26 November 2022, 6:00–7:00).</p>
Full article ">Figure 10
<p>The measurements performed by the beam-crossing sensor installed in location 2 (see <a href="#sensors-25-01393-f004" class="html-fig">Figure 4</a>) on 13 July 2022 (Monday, during a regular semester).</p>
Full article ">Figure 11
<p>Individual durations of beam-crossing events as a function of time and object type (<b>a</b>), and aggregated into histograms for pedestrians, cars, trucks, and buses (<b>b</b>). Based on data from beam-crossing sensor (location “0”) with assigned object type from “Golski” camera sensor, acquired 21–27 November 2022. Events without assigned object type as well as those with durations longer than 2000 ms have been truncated from analysis.</p>
Full article ">Figure 12
<p>Pedestrian traffic intensity on 14 May 2022.</p>
Full article ">Figure 13
<p>Car traffic intensity on 14 May 2022.</p>
Full article ">Figure 14
<p>Directions of movements on the beam-crossing sensor location 3, estimated by camera “Golski”.</p>
Full article ">
15 pages, 3274 KiB  
Article
Gesture-Controlled Robotic Arm for Small Assembly Lines
by Georgios Angelidis and Loukas Bampis
Machines 2025, 13(3), 182; https://doi.org/10.3390/machines13030182 - 25 Feb 2025
Viewed by 218
Abstract
In this study, we present a gesture-controlled robotic arm system for small assembly lines. Robotic arms are extensively used in industrial applications; however, they typically require special treatment and qualified personnel to set up and operate them. Towards this end, hand gestures can [...] Read more.
In this study, we present a gesture-controlled robotic arm system for small assembly lines. Robotic arms are extensively used in industrial applications; however, they typically require special treatment and qualified personnel to set up and operate them. Towards this end, hand gestures can provide a natural way for human–robot interaction, providing a straightforward means for control without the need for significant training of the operators. Our goal is to develop a safe, low-cost, and user-friendly system for environments that often involve non-repetitive and custom automation processes, such as in small factory setups. Our system estimates the 3D position of the user’s joints in real time with the help of AI and real-world data provided by an RGB-D camera. Then, joint coordinates are translated into the robotic arm’s desired poses in a simulated environment (ROS), thus achieving gesture control. Through the experiments we conducted, we show that the system provides the performance required to control a robotic arm effectively and efficiently. Full article
(This article belongs to the Special Issue AI-Integrated Advanced Robotics Towards Industry 5.0)
Show Figures

Figure 1

Figure 1
<p>Schematic representation of our proposed gesture-controlled robotic arm system.</p>
Full article ">Figure 2
<p>Proposed setup for controlling a robotic arm through human gestures. The frames considered for reference are depicted in red for the <span class="html-italic">x</span>-axis, green for the <span class="html-italic">y</span>-axis, and blue for the <span class="html-italic">z</span>-axis.</p>
Full article ">Figure 3
<p>The 21 estimated hand joints from the work presented in [<a href="#B42-machines-13-00182" class="html-bibr">42</a>,<a href="#B50-machines-13-00182" class="html-bibr">50</a>].</p>
Full article ">Figure 4
<p>Direct vectors computed between the index PIP joint and the wrist (blue), pinky MCP, and index MCP joints (red). These vectors are used as references for computing the orientation of the operator’s hand (frame <math display="inline"><semantics> <msub> <mi>h</mi> <mi>r</mi> </msub> </semantics></math>).</p>
Full article ">Figure 5
<p>Rotations of the user’s chest frame of reference (<span class="html-italic">c</span>) in order to align its axes with those of the camera’s coordinate frame (<span class="html-italic">s</span>).</p>
Full article ">Figure 6
<p>The two categories of studied hand poses for controlling the robotic arm. (<b>a</b>) Pose 1, the palm’s surface is perpendicular to the camera; (<b>b</b>) Pose 2, where palm appears parallel relatively to the camera.</p>
Full article ">Figure 7
<p>Snapshots of the developed system’s operation.</p>
Full article ">Figure 8
<p>The two boundary conditions for the gripper’s opening <span class="html-italic">w</span> values.</p>
Full article ">Figure 9
<p>The trajectory of the user’s hand, with respect to the frame of reference <span class="html-italic">c</span>, and the end effector, with respect to <span class="html-italic">b</span>. 33 points are depicted with blue for the human hand and orange for the Panda arm. The left graph is a side-view comparison of the movements, while the right one illustrates the same sequence from a top view.</p>
Full article ">
13 pages, 1987 KiB  
Article
Automatic Detect Incorrect Lifting Posture with the Pose Estimation Model
by Gee-Sern Jison Hsu, Jie Syuan Wu, Yin-Kai Dean Huang, Chun-Chieh Chiu and Jiunn-Horng Kang
Life 2025, 15(3), 358; https://doi.org/10.3390/life15030358 - 24 Feb 2025
Viewed by 245
Abstract
Background: Occupational low back pain (LBP) is a pervasive health issue that significantly impacts productivity and contributes to work-related musculoskeletal disorders (WMSDs). Inadequate lifting postures are a primary, modifiable risk factor associated with LBP, making early detection of unsafe practices crucial to mitigating [...] Read more.
Background: Occupational low back pain (LBP) is a pervasive health issue that significantly impacts productivity and contributes to work-related musculoskeletal disorders (WMSDs). Inadequate lifting postures are a primary, modifiable risk factor associated with LBP, making early detection of unsafe practices crucial to mitigating occupational injuries. Our study aims to address these limitations by developing a markerless, smartphone-based camera system integrated with a deep learning model capable of accurately classifying lifting postures. Material and Method: We recruited 50 healthy adults who participated in lifting tasks using correct and incorrect postures to build a robust dataset. Participants lifted boxes of varying sizes and weights while their movements were recorded from multiple angles and heights to ensure comprehensive data capture. We used the OpenPose algorithm to detect and extract key body points to calculate relevant biomechanical features. These extracted features served as inputs to a bidirectional long short-term memory (LSTM) model, which classified lifting postures into correct and incorrect categories. Results: Our model demonstrated high classification accuracy across all datasets, with accuracy rates of 96.9% for Tr, 95.6% for the testing set, and 94.4% for training. We observed that environmental factors, such as camera angle and height, slightly influenced the model’s accuracy, particularly in scenarios where the subject’s posture partially occluded key body points. Nonetheless, these variations were minor, confirming the robustness of our system across different conditions. Conclusions: This study demonstrates the feasibility and effectiveness of a smartphone camera and AI-based system for lifting posture classification. The system’s high accuracy, low setup cost, and ease of deployment make it a promising tool for enhancing workplace ergonomics. This approach highlights the potential of artificial intelligence to improve occupational safety and underscores the relevance of affordable, scalable solutions in the pursuit of healthier workplaces. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>A</b>) The setting of the recording environment of the study. The camera was placed at three different angles and heights to simulate the conditions of a real-world workplace. (<b>B</b>) The participants were asked to lift the box with correct lifting posture (above row) and incorrect lifting posture according to The National Institute for Occupational Safety and Health (NIOSH) guidelines.</p>
Full article ">Figure 2
<p>The experimental flow protocol.</p>
Full article ">Figure 3
<p>The accuracy and average Euclidean distance of three datasets. Three datasets all have good performance. However, Ty (external validation dataset) has a greater decrease in accuracy and increase in average Euclidean distance, which could be caused by the different video recording conditions.</p>
Full article ">Figure 4
<p>The confusion matrix with heat map of the model. The diagonal grids from upper left to lower right represent the correct predictions of the model. We can find the diagonal grids all have a light color, which means most of the frames are classified into the correct classes (label 0: no life, label 1; lifting with correct posture; label 2: lifting with incorrect posture). The accuracy in classifying correct and incorrect lifting posture is 93.77%, precision is 90.4%, and recall is 95.8%.</p>
Full article ">Figure 5
<p>The accuracy and average Euclidean distance of different camera set. L represents the videos shot at a lower camera height (1 m) and H represents the videos shot at a higher camera height (2.3 m). The accuracy in the H group and coronal group is obviously lower, with a higher average Euclidean distance. The joint angle in the coronal group was greatly affected by 3D to 2D projection. Also, since the camera height is greater than the average human height, the H group has a more severe part occlusion problem compared with the L group.</p>
Full article ">
14 pages, 4858 KiB  
Article
Synthesis and Characterization of Smartphone-Readable Luminescent Lanthanum Borates Doped and Co-Doped with Eu and Dy
by Katya Hristova, Irena P. Kostova, Tinko A. Eftimov, Georgi Patronov and Slava Tsoneva
Photonics 2025, 12(2), 171; https://doi.org/10.3390/photonics12020171 - 19 Feb 2025
Viewed by 371
Abstract
Despite notable advancements in the development of borate materials, improving their luminescent efficiency remains an important focus in materials research. The synthesis of lanthanum borates (LaBO3), doped and co-doped with europium (Eu3⁺) and dysprosium (Dy3⁺), by the [...] Read more.
Despite notable advancements in the development of borate materials, improving their luminescent efficiency remains an important focus in materials research. The synthesis of lanthanum borates (LaBO3), doped and co-doped with europium (Eu3⁺) and dysprosium (Dy3⁺), by the solid-state method, has demonstrated significant potential to address this challenge due to their unique optical properties. These materials facilitate efficient energy transfer from UV-excited host crystals to trivalent rare-earth activators, resulting in stable and high-intensity luminescence. To better understand their structural and vibrational characteristics, Fourier transform infrared (FTIR) spectroscopy and Raman spectroscopy were employed to identify functional groups and molecular vibrations in the synthesized materials. Additionally, X-ray diffraction (XRD) analysis was conducted to determine the crystalline structure and phase composition of the samples. All observed transitions of Eu3⁺ and Dy3⁺ in the excitation and emission spectra were systematically analyzed and identified, providing a comprehensive understanding of their behavior. Although smartphone cameras exhibit non-uniform spectral responses, their integration into this study highlights distinct advantages, including contactless interrogation, effective UV excitation suppression, and real-time spectral analysis. These capabilities enable practical and portable fluorescence sensing solutions for applications in healthcare, environmental monitoring, and food safety. By combining advanced photonic materials with accessible smartphone technology, this work demonstrates a novel approach for developing low-cost, scalable, and innovative sensing platforms that address modern technological demands. Full article
(This article belongs to the Section Optoelectronics and Optical Materials)
Show Figures

Figure 1

Figure 1
<p>Experimental set-up: left is the basic arrangements for the measurement of the 3D excitation–emission spectra of the samples using a standard optical fiber spectrometer; right is a side view of the arrangement from a smartphone camera equipped with a sheet transmission diffraction grating (1000 L/mm) to observe the spectrum.</p>
Full article ">Figure 2
<p>XRD patterns of LaBO<sub>3</sub> doped with Eu and Dy and software-generated crystal structure model of LaBO<sub>3</sub>.</p>
Full article ">Figure 3
<p>FTIR spectra of samples: LaBO<sub>3</sub>:Eu<sup>3+</sup> (S1), LaBO<sub>3</sub>:Dy<sup>3+</sup> (S2), and LaBO<sub>3</sub>:Eu<sup>3+</sup>:Dy<sup>3+</sup> (S3).</p>
Full article ">Figure 4
<p>Raman analysis of sample S3.</p>
Full article ">Figure 5
<p>Raman analysis of samples S1–S3.</p>
Full article ">Figure 6
<p>Graphic representation of the emission intensity at 591 nm, 615 nm, 683 nm, and 708 nm vs. excitation wavelength of 290 nm, 350, 360, 390, 420, 430, 450, and 465 nm for lanthanum borate doped with Eu.</p>
Full article ">Figure 7
<p>Emission spectra of co-doped LaBO<sub>3</sub> at λ<sub>exc</sub> = 290 nm and 350 nm.</p>
Full article ">Figure 8
<p>Topographic and 3D representation of excitation at 290 nm and 396 nm and emission at 589 nm and 615 nm for sample LaBO<sub>3</sub>:Eu<sup>3+</sup>:Dy<sup>3+</sup>, measured by an Ocean Optics spectrometer.</p>
Full article ">Figure 9
<p>Topographic and 3D representation of excitation at 290 nm and 396 nm and emission at 589 nm and 615 nm for sample LaBO<sub>3</sub>:Eu<sup>3+</sup>:Dy<sup>3+</sup>, measured by mobile phone.</p>
Full article ">Figure 10
<p>(<b>a</b>) Comparative graph of samples S1 and S3 for ~615 nm emission at 396 nm excitation, and for sample S2 for 569 nm emission at excitation at 391 nm; (<b>b</b>) comparative graph of samples S1 and S3 at ~591 nm emission at 290 nm and 396 nm excitation, and S2 at 569 nm emission at 393 nm excitation.</p>
Full article ">Scheme 1
<p>Representation of the synthesis of luminescent materials.</p>
Full article ">
16 pages, 2948 KiB  
Article
High-Speed 6DoF Tool Monitoring Using a Low-Cost Photogrammetric System
by Ben Sargeant, Pablo Puerto, Charles Richards, Ibai Leizea, Asier Garcia and Stuart Robson
Metrology 2025, 5(1), 13; https://doi.org/10.3390/metrology5010013 - 19 Feb 2025
Viewed by 264
Abstract
The capability of low-cost photogrammetric systems to provide live six degrees of freedom (6DoF) tracking of multiple objects offers great value in providing high-quality and consistent part production by automated manufacturing systems. However, monitoring of high-speed components, such as cutting heads, presents several [...] Read more.
The capability of low-cost photogrammetric systems to provide live six degrees of freedom (6DoF) tracking of multiple objects offers great value in providing high-quality and consistent part production by automated manufacturing systems. However, monitoring of high-speed components, such as cutting heads, presents several unique challenges that existing systems struggle to meet. The solution given here uses a small number of short-exposure imaging sensors coupled with high-intensity lighting and retrorefective markers to allow for high-speed capture. The use of an initial characterization process carried out using IDEKO’s VSET© system is followed by live object tracking in bespoke image processing software running on a consumer-grade computer. Once this system is in use, it can simultaneously capture images of multiple markers in less than 0.1 milliseconds and use these to determine the 6DoF of the objects that the markers define. 6DoF recalculation of all objects within each measurement instance makes the system resilient to large movements, object occlusion, and sensor relocation. Feasibility tests of a four-camera system as a machine characterization tool tracking a cutting tool spinning at up to 3000 rpm across a volume of 1 m3 achieved a mean reference marker agreement between tool poses of 2.5 µm with markers moving at up to 17.5 ms−1. Given good photogrammetric geometry, 6DoF parameters of the spinning tool were determined to standard deviations of 37.7 µm and 0.086°. Full article
Show Figures

Figure 1

Figure 1
<p>Diagram of the Soraluce multi-axis milling machine with linear axes X, Y, Z and rotational axes A and B. Optionally work space W may also include a rotation table.</p>
Full article ">Figure 2
<p>One of the four tracking system cameras deployed in the milling machine workspace.</p>
Full article ">Figure 3
<p>VMS ‘Domino’ markers on the Soraluce machine head and machine tool.</p>
Full article ">Figure 4
<p>VSET markers across the Soraluce machine for static measurement.</p>
Full article ">Figure 5
<p>Aligned machine tool spindle frames as shown in Spatial Analyzer. Successive 6DoF frames are colored in green and red showing the local Y and X axes spinning about the Z axis in blue. Transmitted at 2 Hz into Spatial Analyzer, the graphic shows progressive sampling of the spinning tool.</p>
Full article ">Figure 6
<p>Example of 6DoF frames from the rotating spindle captured at a single machine pose and showing a number of gross outliers. Contrast to <a href="#metrology-05-00013-f007" class="html-fig">Figure 7</a> where after outlier removal the colour red, green, blue axis coding of the spin about the local Z axis is clear.</p>
Full article ">Figure 7
<p>Sample plots from combined boom and tool spindle 6DoF tracking at tool poses of +/−20 degrees prior to filtering. Jumps in the traces extend from 0.1 mm to 5 mm and represent worst-case non-convergence failures in the 6DoF processing.</p>
Full article ">Figure 8
<p>Example of rotating tool spindle frames captured at a single tool pose after gross outlier removal. Sampled rotations show the local X and Y axes of the tool in red and green respectively spinning about the local Z axis in blue. The small spread visible in the local Z axis is atributable to the deviation of the 6DoF as the tool rotates (<a href="#metrology-05-00013-t001" class="html-table">Table 1</a>).</p>
Full article ">Figure 9
<p>Tool spindle 6DoF spatial and angular standard deviations across nine machine poses after gross outlier removal.</p>
Full article ">Figure 10
<p>Tool spindle 6DoF measurement average standard deviations with different processing workflows and reference marker constellations.</p>
Full article ">
26 pages, 7794 KiB  
Article
Advancing Water Hyacinth Recognition: Integration of Deep Learning and Multispectral Imaging for Precise Identification
by Diego Alberto Herrera Ollachica, Bismark Kweku Asiedu Asante and Hiroki Imamura
Remote Sens. 2025, 17(4), 689; https://doi.org/10.3390/rs17040689 - 18 Feb 2025
Viewed by 313
Abstract
The aquatic plant species Eichhornia crassipes, commonly known as water hyacinth, is indigenous to South America and is considered an invasive species. The invasive water hyacinth has caused significant economic and ecological damage by preventing sunlight from penetrating the surface of the water, [...] Read more.
The aquatic plant species Eichhornia crassipes, commonly known as water hyacinth, is indigenous to South America and is considered an invasive species. The invasive water hyacinth has caused significant economic and ecological damage by preventing sunlight from penetrating the surface of the water, resulting in the loss of aquatic life. To quantify the invasiveness and address the issue of accurately identifying plant species, water hyacinths have prompted numerous researchers to propose approaches to detect regions occupied by water hyacinths. One such solution involves the utilization of multispectral imaging which obtain detailed information about plant species based on the surface reflectance index. This is achieved by analyzing the intensity of light spectra at different wavelengths emitted by each plant. However, the use of multispectral imagery presents a potential challenge since there are various spectral indices that can be used to capture different information. Despite the high accuracy of these multispectral images, there remains a possibility that plants similar to water hyacinths may be misclassified if the right spectral index is not chosen. Considering this challenge, the objective of this research is to develop a low-cost multispectral camera capable of capturing multispectral images. The camera will be equipped with two infrared light spectrum filters with wavelengths of 720 and 850 nanometers, respectively, as well as red, blue, and green light spectrum filters. Additionally, the implementation of the U-Net architecture is proposed for semantic segmentation to accurately identify water hyacinths, as well as other classes such as lakes and land. An accuracy rate of 96% was obtained for the identification of water hyacinths using data captured by an autonomous drone constructed in the laboratory flying at an altitude of 10 m. We also analyzed the contribution each of the infrared layers to the camera’s spectrum setup. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>An illustration by Assman et al. [<a href="#B36-remotesensing-17-00689" class="html-bibr">36</a>] depicting the reflectance maps obtained by a multispectral camera on vegetation.</p>
Full article ">Figure 2
<p>Custom multispectral camera with labeled channels: top-left NIR 850 nm (#133), top-right NIR 720 nm (#67), center RGB (#17), and bottom NoIR (#249).</p>
Full article ">Figure 3
<p>Captured images from the custom-built multispectral camera system: (<b>a</b>) RGB camera, (<b>b</b>) 720 nm infrared camera, (<b>c</b>) 850 nm infrared camera, and (<b>d</b>) NoIR camera.</p>
Full article ">Figure 4
<p>Intersection over union of the four images: (<b>a</b>) RGB camera, (<b>b</b>) 720 nm infrared camera, (<b>c</b>) 850 nm infrared camera, and (<b>d</b>) NoIR camera.</p>
Full article ">Figure 5
<p>Cropped images using the IOU generated after the image processing described in Algorithm 1. (<b>a</b>) RGB camera, (<b>b</b>) 720 nm infrared camera, (<b>c</b>) 850 nm infrared camera, and (<b>d</b>) NoIR camera.</p>
Full article ">Figure 6
<p>Dataset distribution across train, validation, and test subsets for each spectral band.</p>
Full article ">Figure 7
<p>Pixel-wise class distribution in the ground truth masks.</p>
Full article ">Figure 8
<p>U-Net architecture based on a base model with dimensions and parameters per block.</p>
Full article ">Figure 9
<p>NDVI-derived masks.</p>
Full article ">Figure 10
<p>Predicted masks.</p>
Full article ">Figure 11
<p>RGB image.</p>
Full article ">
25 pages, 13237 KiB  
Article
A High-Precision Virtual Central Projection Image Generation Method for an Aerial Dual-Camera
by Xingzhou Luo, Haitao Zhao, Yaping Liu, Nannan Liu, Jiang Chen, Hong Yang and Jie Pan
Remote Sens. 2025, 17(4), 683; https://doi.org/10.3390/rs17040683 - 17 Feb 2025
Viewed by 289
Abstract
Aerial optical cameras are the primary method for capturing high-resolution images to produce large-scale mapping products. To improve aerial photography efficiency, multiple cameras are often used in combination to generate large-format virtual central projection images. This paper presents a high-precision method for directly [...] Read more.
Aerial optical cameras are the primary method for capturing high-resolution images to produce large-scale mapping products. To improve aerial photography efficiency, multiple cameras are often used in combination to generate large-format virtual central projection images. This paper presents a high-precision method for directly transforming raw images obtained from a dual-camera system mounted at an oblique angle into virtual central projection images, thereby enabling the construction of low-cost, large-format aerial camera systems. The method commences with an adaptive sub-block in the overlapping regions of the raw images to extract evenly distributed feature points, followed by iterative relative orientation to improve accuracy and reliability. A global projection transformation matrix is constructed, and the sigmoid function is employed as a weighted distance function for image stitching. The results demonstrate that the proposed method produces more evenly distributed feature points, higher relative orientation accuracy, and greater reliability. Simulation analysis of image overlap indicates that when the overlap exceeds 7%, stitching accuracy can be better than 1.25 μm. The aerial triangulation results demonstrate that the virtual central projection images satisfy the criteria for the production of 1:1000 scale mapping products. Full article
Show Figures

Figure 1

Figure 1
<p>Imaging principle of ALC 2000. (<b>a</b>) Design model. (<b>b</b>) Geometry model.</p>
Full article ">Figure 2
<p>Workflow of virtual image generation.</p>
Full article ">Figure 3
<p>Diagram of adaptive sub-block partitioning with pre-transformation in overlapping region.</p>
Full article ">Figure 4
<p>Flowchart of feature point extraction and matching algorithm.</p>
Full article ">Figure 5
<p>Flowchart of relative orientation calibration algorithm.</p>
Full article ">Figure 6
<p>Distance weight curve of overlapping region.</p>
Full article ">Figure 7
<p>Coverage range of experimental flight routes in Hefei and typical data.</p>
Full article ">Figure 8
<p>SIFT based on adaptive sub-block partitioning with pre-transformation: (<b>a</b>) city; (<b>b</b>) forest; and (<b>c</b>) farmland.</p>
Full article ">Figure 9
<p>Distribution and point density analysis of corresponding feature points in overlapping region using different methods: (<b>a</b>) Direct SIFT extraction in the overlap. (<b>b</b>) Direct SURF extraction in the overlap. (<b>c</b>) Direct AKAZE extraction in the overlap. (<b>d</b>) Direct ORB extraction in the overlap. (<b>e</b>) LoFTR extraction based on adaptive sub-block partitioning with pre-transformation. (<b>f</b>) SIFT extraction based on adaptive sub-block partitioning with pre-transformation (ASPPT-SIFT, ours).</p>
Full article ">Figure 10
<p>The relative orientation calibration results: (<b>a</b>) The relative orientation angle (φ) result. (<b>b</b>) The relative orientation angle (ω) result. (<b>c</b>) The relative orientation angle (κ) result.</p>
Full article ">Figure 11
<p>The relative orientation accuracy assessment results: (<b>a</b>) The relative orientation angle (φ) accuracy. (<b>b</b>) The relative orientation angle (ω) accuracy. (<b>c</b>) The relative orientation angle (κ) accuracy.</p>
Full article ">Figure 11 Cont.
<p>The relative orientation accuracy assessment results: (<b>a</b>) The relative orientation angle (φ) accuracy. (<b>b</b>) The relative orientation angle (ω) accuracy. (<b>c</b>) The relative orientation angle (κ) accuracy.</p>
Full article ">Figure 12
<p>Disparity distance root mean square error.</p>
Full article ">Figure 13
<p>Simulation data of different image overlaps.</p>
Full article ">Figure 14
<p>Relative orientation calibration accuracy of different overlaps.</p>
Full article ">
Back to TopTop