[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (502)

Search Parameters:
Keywords = odometry

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 3529 KiB  
Article
Meta-Transfer-Learning-Based Multimodal Human Pose Estimation for Lower Limbs
by Guoming Du, Haiqi Zhu, Zhen Ding, Hong Huang, Xiaofeng Bie and Feng Jiang
Sensors 2025, 25(5), 1613; https://doi.org/10.3390/s25051613 - 6 Mar 2025
Abstract
Accurate and reliable human pose estimation (HPE) is essential in interactive systems, particularly for applications requiring personalized adaptation, such as controlling cooperative robots and wearable exoskeletons, especially for healthcare monitoring equipment. However, continuously maintaining diverse datasets and frequently updating models for individual adaptation [...] Read more.
Accurate and reliable human pose estimation (HPE) is essential in interactive systems, particularly for applications requiring personalized adaptation, such as controlling cooperative robots and wearable exoskeletons, especially for healthcare monitoring equipment. However, continuously maintaining diverse datasets and frequently updating models for individual adaptation are both resource intensive and time-consuming. To address these challenges, we propose a meta-transfer learning framework that integrates multimodal inputs, including high-frequency surface electromyography (sEMG), visual-inertial odometry (VIO), and high-precision image data. This framework improves both accuracy and stability through a knowledge fusion strategy, resolving the data alignment issue, ensuring seamless integration of different modalities. To further enhance adaptability, we introduce a training and adaptation framework with few-shot learning, facilitating efficient updating of encoders and decoders for dynamic feature adjustment in real-time applications. Experimental results demonstrate that our framework provides accurate, high-frequency pose estimations, particularly for intra-subject adaptation. Our approach enables efficient adaptation to new individuals with only a few new samples, providing an effective solution for personalized motion analysis with minimal data. Full article
Show Figures

Figure 1

Figure 1
<p>Sensor placement and data collection environment: (<b>a</b>) For the lower body, six sEMG sensors were placed on both sides of the legs, while 16 Vicon markers were used to collect ground-truth data. An Intel RealSense T265 sensor was mounted on the waist. (<b>b</b>) Ten Vicon cameras were positioned on the ceiling to capture reflective markers on the lower body, and an RGB camera was placed on the side wall. The subject performed walking trials on flat ground, both clockwise and counterclockwise.</p>
Full article ">Figure 2
<p>Overall schematic of proposed framework, totally including three phases.</p>
Full article ">Figure 3
<p>The pose estimation network is pipelined with feature extraction, knowledge sharing, fusion of knowledge and pose regression.</p>
Full article ">Figure 4
<p>The structure of CBAM-Resnet12 is composed of a combination of CBAM module, residual block, convolution layer and max pooling layer.</p>
Full article ">Figure 5
<p>Results on different subjects with different scales of pre-training.</p>
Full article ">Figure 6
<p>Evaluation of different joints from lower body, results are calculated with RMSE in degrees.</p>
Full article ">
28 pages, 4077 KiB  
Review
A Comprehensive Survey on Short-Distance Localization of UAVs
by Luka Kramarić, Niko Jelušić, Tomislav Radišić and Mario Muštra
Drones 2025, 9(3), 188; https://doi.org/10.3390/drones9030188 - 4 Mar 2025
Viewed by 145
Abstract
The localization of Unmanned Aerial Vehicles (UAVs) is a critical area of research, particularly in applications requiring high accuracy and reliability in Global Positioning System (GPS)-denied environments. This paper presents a comprehensive overview of short-distance localization methods for UAVs, exploring their strengths, limitations, [...] Read more.
The localization of Unmanned Aerial Vehicles (UAVs) is a critical area of research, particularly in applications requiring high accuracy and reliability in Global Positioning System (GPS)-denied environments. This paper presents a comprehensive overview of short-distance localization methods for UAVs, exploring their strengths, limitations, and practical applications. Among short-distance localization methods, ultra-wideband (UWB) technology has gained significant attention due to its ability to provide accurate positioning, resistance to multipath interference, and low power consumption. Different approaches to the usage of UWB sensors, such as time of arrival (ToA), time difference of arrival (TDoA), and double-sided two-way ranging (DS-TWR), alongside their integration with complementary sensors like Inertial Measurement Units (IMUs), cameras, and visual odometry systems, are explored. Furthermore, this paper provides an evaluation of the key factors affecting UWB-based localization performance, including anchor placement, synchronization, and the challenges of combined use with other localization technologies. By highlighting the current trends in UWB-related research, including its increasing use in swarm control, indoor navigation, and autonomous landing, potential researchers could benefit from this study by using it as a guide for choosing the appropriate localization techniques, emphasizing UWB technology’s potential as a foundational technology in advanced UAV applications. Full article
(This article belongs to the Special Issue Resilient Networking and Task Allocation for Drone Swarms)
Show Figures

Figure 1

Figure 1
<p>The steps in the process of designing a short-distance localization system for UAVs, from the choice of the application and the environment to the required performance.</p>
Full article ">Figure 2
<p>The principle of the Extended Kalman Filter allows the usage of a linear filter in nonlinear state estimation [<a href="#B19-drones-09-00188" class="html-bibr">19</a>].</p>
Full article ">Figure 3
<p>The localization trajectories of a UAV where the yellow, cyan, red, and blue curves represent the ground truth, UWB, QVIO, and AprilTag, respectively [<a href="#B35-drones-09-00188" class="html-bibr">35</a>].</p>
Full article ">Figure 4
<p>Landing locations by using different localization equipment [<a href="#B57-drones-09-00188" class="html-bibr">57</a>].</p>
Full article ">Figure 5
<p>The localization error with and without the use of UWB with GNSS/IMU shows that the combination of localization systems provides significantly better results [<a href="#B69-drones-09-00188" class="html-bibr">69</a>].</p>
Full article ">Figure 6
<p>A comparison of the trajectories from different combinations of aids to UWB technology: (<b>a</b>) Integration between the camera and INS for 180 s of a complete signal outage; (<b>b</b>) INS dead reckoning solution compared against reference trajectory for 60 s of GNSS signals outage; and (<b>c</b>) UWB-INS integration performance compared against reference trajectory for 180 s GNSS signal outage [<a href="#B73-drones-09-00188" class="html-bibr">73</a>].</p>
Full article ">Figure 7
<p>Message exchange for a single UAV-anchor pair using the DS-TWR [<a href="#B74-drones-09-00188" class="html-bibr">74</a>].</p>
Full article ">Figure 8
<p>The real vs. the predefined flight trajectory in the xy-plane [<a href="#B74-drones-09-00188" class="html-bibr">74</a>].</p>
Full article ">
19 pages, 1958 KiB  
Article
Visual-Inertial-Wheel Odometry with Slip Compensation and Dynamic Feature Elimination
by Niraj Reginald, Omar Al-Buraiki, Thanacha Choopojcharoen, Baris Fidan and Ehsan Hashemi
Sensors 2025, 25(5), 1537; https://doi.org/10.3390/s25051537 - 1 Mar 2025
Viewed by 210
Abstract
Inertial navigation systems augmented with visual and wheel odometry measurements have emerged as a robust solution to address uncertainties in robot localization and odometry. This paper introduces a novel data-driven approach to compensate for wheel slippage in visual-inertial-wheel odometry (VIWO). The proposed method [...] Read more.
Inertial navigation systems augmented with visual and wheel odometry measurements have emerged as a robust solution to address uncertainties in robot localization and odometry. This paper introduces a novel data-driven approach to compensate for wheel slippage in visual-inertial-wheel odometry (VIWO). The proposed method leverages Gaussian process regression (GPR) with deep kernel design and long short-term memory (LSTM) layers to model and mitigate slippage-induced errors effectively. Furthermore, a feature confidence estimator is incorporated to address the impact of dynamic feature points on visual measurements, ensuring reliable data integration. By refining these measurements, the system utilizes a multi-state constraint Kalman filter (MSCKF) to achieve accurate state estimation and enhanced navigation performance. The effectiveness of the proposed approach is demonstrated through extensive simulations and experimental validations using real-world datasets. The results highlight the ability of the method to handle challenging terrains and dynamic environments by compensating for wheel slippage and mitigating the influence of dynamic objects. Compared to conventional VIWO systems, the integration of GPR and LSTM layers significantly improves localization accuracy and robustness. This work paves the way for deploying VIWO systems in diverse and unpredictable environments, contributing to advancements in autonomous navigation and multi-sensor fusion technologies. Full article
(This article belongs to the Section Physical Sensors)
Show Figures

Figure 1

Figure 1
<p>Overall structure of the proposed VIWO scheme.</p>
Full article ">Figure 2
<p>Slip compensation scheme.</p>
Full article ">Figure 3
<p>Robot kinematic schematic.</p>
Full article ">Figure 4
<p>Deep kernel architecture using CNN.</p>
Full article ">Figure 5
<p>Test sequence 1 odometry trajectories comparison. VIWO: visual-inertial-wheel odometry, DPE: dynamic point elimination, WSC with CNN: wheel-slip compensation with CNN for kernel design, WSC with RNN: wheel-slip compensation with RNN for kernel design.</p>
Full article ">Figure 6
<p>Test sequence 2 odometry trajectories comparison.</p>
Full article ">Figure 7
<p>MATE error for test sequences.</p>
Full article ">Figure 8
<p>Experimental setup.</p>
Full article ">Figure 9
<p>Comparison of wheel speeds of the robot.</p>
Full article ">Figure 10
<p>Experiment sequence 1 odometry trajectory comparison.</p>
Full article ">Figure 11
<p>Experiment sequence 2 odometry trajectory comparison.</p>
Full article ">Figure 12
<p>RMSE for the experimental trajectories.</p>
Full article ">Figure 13
<p>Experiment sequence 3 odometry trajectories comparison in gravel/sand-based terrain.</p>
Full article ">
21 pages, 10896 KiB  
Article
Loosely Coupled PPP/Inertial/LiDAR Simultaneous Localization and Mapping (SLAM) Based on Graph Optimization
by Baoxiang Zhang, Cheng Yang, Guorui Xiao, Peigong Li, Zhengyang Xiao, Haopeng Wei and Jialin Liu
Remote Sens. 2025, 17(5), 812; https://doi.org/10.3390/rs17050812 - 25 Feb 2025
Viewed by 193
Abstract
Navigation services and high-precision positioning play a significant role in emerging fields such as self-driving and mobile robots. The performance of precise point positioning (PPP) may be seriously affected by signal interference and struggles to achieve continuous and accurate positioning in complex environments. [...] Read more.
Navigation services and high-precision positioning play a significant role in emerging fields such as self-driving and mobile robots. The performance of precise point positioning (PPP) may be seriously affected by signal interference and struggles to achieve continuous and accurate positioning in complex environments. LiDAR/inertial navigation can use spatial structure information to realize pose estimation but cannot solve the problem of cumulative error. This study proposes a PPP/inertial/LiDAR combined localization algorithm based on factor graph optimization. Firstly, the algorithm performed the spatial alignment by adding the initial yaw factor. Then, the PPP factor and anchor factor were constructed using PPP information. Finally, the global localization is estimated accurately and robustly based on the factor graph. The vehicle experiment shows that the proposed algorithm in this study can achieve meter-level accuracy in complex environments and can greatly enhance the accuracy, continuity, and reliability of attitude estimation. Full article
Show Figures

Figure 1

Figure 1
<p>The diagram illustrates the workflow of the proposed system. Initially, the measurements from all sensors are preprocessed. During the initialization phase, LiDAR inertial initialization is performed by aligning the inertial data with LiDAR odometry. Next, PPP is synchronized with LIO through initial yaw optimization. Finally, the measurement constraints are refined through nonlinear optimization.</p>
Full article ">Figure 2
<p>The optimization problem is represented as a factor graph, where the system states are depicted by large yellow circles, and the factors are shown as small colored circles.</p>
Full article ">Figure 3
<p>Distribution of sensors in the data acquisition platform.</p>
Full article ">Figure 4
<p>Experiment 1 vehicle trajectory with the number of satellites.</p>
Full article ">Figure 5
<p>Experiment 2 vehicle trajectory with the number of satellites.</p>
Full article ">Figure 6
<p>The initial yaw estimation errors in Experiment 1.</p>
Full article ">Figure 7
<p>The initial yaw estimation errors in Experiment 2.</p>
Full article ">Figure 8
<p>The errors of anchor point optimization in Experiment 1.</p>
Full article ">Figure 9
<p>The errors of anchor point optimization in Experiment 2.</p>
Full article ">Figure 10
<p>The NAST and the PDOP in Experiment 1.</p>
Full article ">Figure 11
<p>The NAST and the PDOP in Experiment 2.</p>
Full article ">Figure 12
<p>Comparison of position accuracy of two algorithms in Experiment 1.</p>
Full article ">Figure 13
<p>Comparison of position accuracy of two algorithms in Experiment 2.</p>
Full article ">Figure 14
<p>Experiment 1: comparison of attitude accuracy of two algorithms.</p>
Full article ">Figure 15
<p>Experiment 2: comparison of attitude accuracy of two algorithms.</p>
Full article ">Figure 16
<p>Experiment 1: LIO, PPP position increment, and LIOP error curves.</p>
Full article ">Figure 17
<p>LIO, PPP position increment, and LIOP error curves in Experiment 2.</p>
Full article ">Figure 18
<p>Position increment analysis: (<b>a</b>) enlarged view of the Inc E curve from 170 s to 270 s in <a href="#remotesensing-17-00812-f017" class="html-fig">Figure 17</a>; (<b>b</b>) enlarged view of the Inc E curve from 410 s to 510 s in <a href="#remotesensing-17-00812-f017" class="html-fig">Figure 17</a>.</p>
Full article ">
21 pages, 901 KiB  
Article
Multi-Sensor Information Fusion for Mobile Robot Indoor-Outdoor Localization: A Zonotopic Set-Membership Estimation Approach
by Yanfei Zhu, Xuanyu Fang and Chuanjiang Li
Electronics 2025, 14(5), 867; https://doi.org/10.3390/electronics14050867 - 21 Feb 2025
Viewed by 394
Abstract
This paper investigates the localization of mobile robots in both indoor and outdoor scenarios. A zonotopic set-membership approach is proposed to fuse global navigation satellite system and odometry data outdoors, and 2D laser and odometry data indoors. Seamless switching between indoor and outdoor [...] Read more.
This paper investigates the localization of mobile robots in both indoor and outdoor scenarios. A zonotopic set-membership approach is proposed to fuse global navigation satellite system and odometry data outdoors, and 2D laser and odometry data indoors. Seamless switching between indoor and outdoor scene localization is achieved through a comparison of the current global navigation satellite system signal’s covariance with a predefined threshold in the proposed approach. Firstly, the robot’s position information is characterized using the odometry model, and the set containing the true state of the robot is updated to obtain the current updated zonotope. In addition, the global navigation satellite system or laser observation equations are described as a strip region and intersected with the prediction zonotope to obtain the feasible set of the states. Choosing the zonotope with the smallest volume from a family that encompasses the intersection of the two serves as the outer boundary for the intersection, enabling the determination of the precise position. The algorithm proposed in this paper can estimate the position state of the mobile robot to achieve accurate localization. To validate the proposed approach, relevant data are presented in the simulation results and discussion. Full article
(This article belongs to the Special Issue Multisensor Fusion: Latest Advances and Prospects)
Show Figures

Figure 1

Figure 1
<p>The odometry model of mobile robot.</p>
Full article ">Figure 2
<p>Outdoor position state estimation.</p>
Full article ">Figure 3
<p>Indoor position state estimation.</p>
Full article ">Figure 4
<p>Outdoor simulation trajectory of the mobile robot.</p>
Full article ">Figure 5
<p>Indoor simulation trajectory of the mobile robot.</p>
Full article ">Figure 6
<p>Indoor and outdoor switching positioning trajectory diagram.</p>
Full article ">Figure 7
<p>Outdoor position state estimation in <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>1</mn> <mo>:</mo> <mn>16</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 8
<p>Indoor position state estimation in <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>17</mn> <mo>:</mo> <mn>46</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 9
<p>Outdoor position state estimation in <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>=</mo> <mn>47</mn> <mo>:</mo> <mn>70</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 10
<p>Estimated values of the two algorithm under Case 1.</p>
Full article ">Figure 11
<p>Estimated values of the two algorithms under Case 2.</p>
Full article ">Figure 12
<p>Estimated values of the two algorithms under Case 3.</p>
Full article ">Figure 12 Cont.
<p>Estimated values of the two algorithms under Case 3.</p>
Full article ">Figure 13
<p>The positioning trajectory diagram under three cases.</p>
Full article ">Figure 13 Cont.
<p>The positioning trajectory diagram under three cases.</p>
Full article ">
15 pages, 3120 KiB  
Article
Implementation of Visual Odometry on Jetson Nano
by Jakub Krško, Dušan Nemec, Vojtech Šimák and Mário Michálik
Sensors 2025, 25(4), 1025; https://doi.org/10.3390/s25041025 - 9 Feb 2025
Viewed by 571
Abstract
This paper presents the implementation of ORB-SLAM3 for visual odometry on a low-power ARM-based system, specifically the Jetson Nano, to track a robot’s movement using RGB-D cameras. Key challenges addressed include the selection of compatible software libraries, camera calibration, and system optimization. The [...] Read more.
This paper presents the implementation of ORB-SLAM3 for visual odometry on a low-power ARM-based system, specifically the Jetson Nano, to track a robot’s movement using RGB-D cameras. Key challenges addressed include the selection of compatible software libraries, camera calibration, and system optimization. The ORB-SLAM3 algorithm was adapted for the ARM architecture and tested using both the EuRoC dataset and real-world scenarios involving a mobile robot. The testing demonstrated that ORB-SLAM3 provides accurate localization, with errors in path estimation ranging from 3 to 11 cm when using the EuRoC dataset. Real-world tests on a mobile robot revealed discrepancies primarily due to encoder drift and environmental factors such as lighting and texture. The paper discusses strategies for mitigating these errors, including enhanced calibration and the potential use of encoder data for tracking when camera performance falters. Future improvements focus on refining the calibration process, adding trajectory correction mechanisms, and integrating visual odometry data more effectively into broader systems. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

Figure 1
<p>Chessboard pattern with highlighted corners [<a href="#B1-sensors-25-01025" class="html-bibr">1</a>].</p>
Full article ">Figure 2
<p>EuRoC dataset sample [<a href="#B20-sensors-25-01025" class="html-bibr">20</a>].</p>
Full article ">Figure 3
<p>Detail of difference between estimated and ground truth trajectories [<a href="#B1-sensors-25-01025" class="html-bibr">1</a>].</p>
Full article ">Figure 4
<p>Comparison between robot and camera trajectories [<a href="#B1-sensors-25-01025" class="html-bibr">1</a>].</p>
Full article ">Figure 5
<p>ORB-SLAM3 GUI [<a href="#B1-sensors-25-01025" class="html-bibr">1</a>].</p>
Full article ">Figure 6
<p>State-chart for implemented algorithm [<a href="#B1-sensors-25-01025" class="html-bibr">1</a>].</p>
Full article ">
29 pages, 4682 KiB  
Article
LSAF-LSTM-Based Self-Adaptive Multi-Sensor Fusion for Robust UAV State Estimation in Challenging Environments
by Mahammad Irfan, Sagar Dalai, Petar Trslic, James Riordan and Gerard Dooly
Machines 2025, 13(2), 130; https://doi.org/10.3390/machines13020130 - 9 Feb 2025
Viewed by 584
Abstract
Unmanned aerial vehicle (UAV) state estimation is fundamental across applications like robot navigation, autonomous driving, virtual reality (VR), and augmented reality (AR). This research highlights the critical role of robust state estimation in ensuring safe and efficient autonomous UAV navigation, particularly in challenging [...] Read more.
Unmanned aerial vehicle (UAV) state estimation is fundamental across applications like robot navigation, autonomous driving, virtual reality (VR), and augmented reality (AR). This research highlights the critical role of robust state estimation in ensuring safe and efficient autonomous UAV navigation, particularly in challenging environments. We propose a deep learning-based adaptive sensor fusion framework for UAV state estimation, integrating multi-sensor data from stereo cameras, an IMU, two 3D LiDAR’s, and GPS. The framework dynamically adjusts fusion weights in real time using a long short-term memory (LSTM) model, enhancing robustness under diverse conditions such as illumination changes, structureless environments, degraded GPS signals, or complete signal loss where traditional single-sensor SLAM methods often fail. Validated on an in-house integrated UAV platform and evaluated against high-precision RTK ground truth, the algorithm incorporates deep learning-predicted fusion weights into an optimization-based odometry pipeline. The system delivers robust, consistent, and accurate state estimation, outperforming state-of-the-art techniques. Experimental results demonstrate its adaptability and effectiveness across challenging scenarios, showcasing significant advancements in UAV autonomy and reliability through the synergistic integration of deep learning and sensor fusion. Full article
Show Figures

Figure 1

Figure 1
<p>Proposed architecture for LSTM-based self-adaptive multi-sensor fusion (LSAF).</p>
Full article ">Figure 2
<p>An illustration of the proposed LSAF framework. The global estimator combines local estimations from various global sensors to achieve precise local accuracy and globally drift free pose estimation, which builds upon our previous work [<a href="#B28-machines-13-00130" class="html-bibr">28</a>].</p>
Full article ">Figure 3
<p>Proposed LSTM-based multi-sensor fusion architecture for UAV state estimation.</p>
Full article ">Figure 4
<p>LSTM cell architecture for adaptive multi-sensor fusion.</p>
Full article ">Figure 5
<p>Training and validation loss of the proposed LSTM-based self-adaptive multi-sensor fusion (LSAF) framework over 1000 epochs.</p>
Full article ">Figure 6
<p>Training and validation MAE of the proposed LSTM-based self-adaptive multi-sensor fusion (LSAF) framework over 1000 epochs.</p>
Full article ">Figure 7
<p>Proposed block diagram for LSTM-based self-adaptive multi-sensor fusion (LSAF).</p>
Full article ">Figure 8
<p>The experimental environment in different scenarios during the data collection. Panel (<b>a</b>,<b>b</b>) represent the UAV hardware along with sensor integration and panel (<b>c</b>,<b>d</b>) are the open-field dataset environment view from stereo and LiDAR sensors, respectively, which build upon our previous work [<a href="#B28-machines-13-00130" class="html-bibr">28</a>].</p>
Full article ">Figure 9
<p>Trajectory plots of the proposed LSAF method and comparison with FASTLIO2 and VINS-Fusion.</p>
Full article ">Figure 10
<p>Box plots showing the overall APE of each strategy.</p>
Full article ">Figure 11
<p>Absolute estimated position of x, y, and z axes showing plots of various methods on the UAV car parking dataset.</p>
Full article ">Figure 12
<p>Absolute position error of roll, yaw, and pitch showing the plots of various methods on the UAV car parking dataset.</p>
Full article ">Figure 13
<p>Trajectory plots of the proposed LSAF method and comparison with FASTLIO2 and VINS-Fusion on the UL outdoor handheld dataset.</p>
Full article ">Figure 14
<p>Box plots showing the overall APE of each strategy.</p>
Full article ">Figure 15
<p>Absolute estimated position of x, y, and z axes showing the plots of various methods on the UL outdoor handheld dataset.</p>
Full article ">Figure 16
<p>Absolute position error of roll, yaw, and pitch showing the plots of various methods on the UL outdoor handheld dataset.</p>
Full article ">Figure 17
<p>Trajectory plots of the proposed LSAF method and comparison with FASTLIO2 and VINS-Fusion.</p>
Full article ">Figure 18
<p>Absolute estimated position of the x, y, and z axes showing the plots of various methods on the UAV car bridge dataset.</p>
Full article ">Figure 19
<p>Absolute position error of roll, yaw, and pitch showing plots of various methods on the UAV car bridge dataset.</p>
Full article ">Figure 20
<p>Box plots showing the overall APE of each strategy.</p>
Full article ">
16 pages, 6121 KiB  
Article
Stereo Event-Based Visual–Inertial Odometry
by Kunfeng Wang, Kaichun Zhao, Wenshuai Lu and Zheng You
Sensors 2025, 25(3), 887; https://doi.org/10.3390/s25030887 - 31 Jan 2025
Cited by 1 | Viewed by 502
Abstract
Event-based cameras are a new type of vision sensor in which pixels operate independently and respond asynchronously to changes in brightness with microsecond resolution, instead of providing standard intensity frames. Compared with traditional cameras, event-based cameras have low latency, no motion blur, and [...] Read more.
Event-based cameras are a new type of vision sensor in which pixels operate independently and respond asynchronously to changes in brightness with microsecond resolution, instead of providing standard intensity frames. Compared with traditional cameras, event-based cameras have low latency, no motion blur, and high dynamic range (HDR), which provide possibilities for robots to deal with some challenging scenes. We propose a visual–inertial odometry for stereo event-based cameras based on Error-State Kalman Filter (ESKF). The vision module updates the pose by relying on the edge alignment of a semi-dense 3D map to a 2D image, while the IMU module updates the pose using median integration. We evaluate our method on public datasets with general 6-DoF motion (three-axis translation and three-axis rotation) and compare the results against the ground truth. We compared our results with those from other methods, demonstrating the effectiveness of our approach. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Figure 1
<p>(<b>Top left</b>): scene. (<b>Bottom left</b>): inverse depth map at time <span class="html-italic">t</span>, and different colors represent different depths. (<b>Right</b>): global map and pose estimation.</p>
Full article ">Figure 2
<p>Overview of our proposed stereo event-based visual–inertial odometry.</p>
Full article ">Figure 3
<p>Time-surface. (<b>Left</b>): output of an event camera, and different colors represent different times. (<b>Right</b>): time-surface map. Figure adapted from [<a href="#B16-sensors-25-00887" class="html-bibr">16</a>].</p>
Full article ">Figure 4
<p>Time-surface and its included historical information.</p>
Full article ">Figure 5
<p>Algorithm performance. The left image shows the experimental scene, while the right image displays the local point clouds and trajectories.</p>
Full article ">Figure 6
<p>The first column shows images from a traditional camera. The second column is the time-surface. The third column is the inverse depth map. The last column is the warping depth map overlaid on the time-surface negative.</p>
Full article ">
42 pages, 40649 KiB  
Article
A Multi-Drone System Proof of Concept for Forestry Applications
by André G. Araújo, Carlos A. P. Pizzino, Micael S. Couceiro and Rui P. Rocha
Drones 2025, 9(2), 80; https://doi.org/10.3390/drones9020080 - 21 Jan 2025
Viewed by 1262
Abstract
This study presents a multi-drone proof of concept for efficient forest mapping and autonomous operation, framed within the context of the OPENSWARM EU Project. The approach leverages state-of-the-art open-source simultaneous localisation and mapping (SLAM) frameworks, like LiDAR (Light Detection And Ranging) Inertial Odometry [...] Read more.
This study presents a multi-drone proof of concept for efficient forest mapping and autonomous operation, framed within the context of the OPENSWARM EU Project. The approach leverages state-of-the-art open-source simultaneous localisation and mapping (SLAM) frameworks, like LiDAR (Light Detection And Ranging) Inertial Odometry via Smoothing and Mapping (LIO-SAM), and Distributed Collaborative LiDAR SLAM Framework for a Robotic Swarm (DCL-SLAM), seamlessly integrated within the MRS UAV System and Swarm Formation packages. This integration is achieved through a series of procedures compliant with Robot Operating System middleware (ROS), including an auto-tuning particle swarm optimisation method for enhanced flight control and stabilisation, which is crucial for autonomous operation in challenging environments. Field experiments conducted in a forest with multiple drones demonstrate the system’s ability to navigate complex terrains as a coordinated swarm, accurately and collaboratively mapping forest areas. Results highlight the potential of this proof of concept, contributing to the development of scalable autonomous solutions for forestry management. The findings emphasise the significance of integrating multiple open-source technologies to advance sustainable forestry practices using swarms of drones. Full article
Show Figures

Figure 1

Figure 1
<p>System architecture proposed for the multi-drone PoC system.</p>
Full article ">Figure 2
<p>The world frame <math display="inline"><semantics> <mrow> <mi>w</mi> <mo>=</mo> <mfenced separators="" open="{" close="}"> <msub> <mi mathvariant="bold">e</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi mathvariant="bold">e</mi> <mn>2</mn> </msub> <mo>,</mo> <msub> <mi mathvariant="bold">e</mi> <mn>3</mn> </msub> </mfenced> </mrow> </semantics></math>, in which the position and orientation of the drone is expressed by translation <math display="inline"><semantics> <mrow> <mi mathvariant="bold">r</mi> <mo>=</mo> <msup> <mrow> <mo>[</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>,</mo> <mi>z</mi> <mo>]</mo> </mrow> <mi>T</mi> </msup> </mrow> </semantics></math> and rotation <math display="inline"><semantics> <mrow> <mi mathvariant="bold">R</mi> <mo>(</mo> <mi>ϕ</mi> <mo>,</mo> <mi>θ</mi> <mo>,</mo> <mi>ψ</mi> <mo>)</mo> </mrow> </semantics></math> to the body frame <math display="inline"><semantics> <mrow> <mi>b</mi> <mo>=</mo> <mfenced separators="" open="{" close="}"> <msub> <mi mathvariant="bold">b</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi mathvariant="bold">b</mi> <mn>2</mn> </msub> <mo>,</mo> <msub> <mi mathvariant="bold">b</mi> <mn>3</mn> </msub> </mfenced> </mrow> </semantics></math>. The drone heading vector <math display="inline"><semantics> <mi mathvariant="bold">h</mi> </semantics></math>, which is a projection of <math display="inline"><semantics> <msub> <mover accent="true"> <mi mathvariant="bold">b</mi> <mo stretchy="false">^</mo> </mover> <mn>1</mn> </msub> </semantics></math> to the plane <math display="inline"><semantics> <mrow> <mo form="prefix">span</mo> <mfenced separators="" open="(" close=")"> <msub> <mover accent="true"> <mi mathvariant="bold">e</mi> <mo stretchy="false">^</mo> </mover> <mn>1</mn> </msub> <mo>,</mo> <msub> <mover accent="true"> <mi mathvariant="bold">e</mi> <mo stretchy="false">^</mo> </mover> <mn>2</mn> </msub> </mfenced> </mrow> </semantics></math>, forms the heading angle <math display="inline"><semantics> <mrow> <mi>η</mi> <mo>=</mo> <mo form="prefix">atan</mo> <mn>2</mn> <mfenced separators="" open="(" close=")"> <msubsup> <mover accent="true"> <mi mathvariant="bold">b</mi> <mo stretchy="false">^</mo> </mover> <mn>1</mn> <mo>⊤</mo> </msubsup> <msub> <mover accent="true"> <mi mathvariant="bold">e</mi> <mo stretchy="false">^</mo> </mover> <mn>2</mn> </msub> <mo>,</mo> <msubsup> <mover accent="true"> <mi mathvariant="bold">b</mi> <mo stretchy="false">^</mo> </mover> <mn>1</mn> <mo>⊤</mo> </msubsup> <msub> <mover accent="true"> <mi mathvariant="bold">e</mi> <mo stretchy="false">^</mo> </mover> <mn>1</mn> </msub> </mfenced> <mo>=</mo> <mo form="prefix">atan</mo> <mn>2</mn> <mfenced separators="" open="(" close=")"> <msub> <mi mathvariant="bold">h</mi> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> </msub> <mo>,</mo> <msub> <mi mathvariant="bold">h</mi> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </msub> </mfenced> </mrow> </semantics></math>, figure based on [<a href="#B6-drones-09-00080" class="html-bibr">6</a>].</p>
Full article ">Figure 3
<p>Figure based on [<a href="#B6-drones-09-00080" class="html-bibr">6</a>]. The filters simultaneously estimate the states and can be switched or selected by user/arbiter.</p>
Full article ">Figure 4
<p>Simulation of the swarm formation in the forest environment. Together, these visualisations demonstrate the effectiveness of the simulation tools in evaluating and refining the Multi-Drone PoC system prior to field experiments. (<b>a</b>) Octomap representation of a simulated forest environment in Gazebo, shown using a color gradient that varies with height. (<b>b</b>) Representation of swarm formation in the simulation environment. The three colors (pink, green, and blue) in small dot points represent the global maps of each drone. The square markers indicate the reference samples from the Octomap planner’s desired trajectory. The trajectory, represented by vectors, corresponds to the outputs of the MPC tracker. Additionally, the actual paths of each drone are depicted as solid lines. Finally, the solid red lines represent the current swarm formation shape.</p>
Full article ">Figure 5
<p>Global map service. (<b>a</b>) Overview of the global map integration process, where local maps from each drone are collected and aligned using the Iterative Closest Point (ICP) algorithm to create a unified global map of the environment. (<b>b</b>) Resulting integrated global map generated by combining the local maps from three drones, namely drone <math display="inline"><semantics> <mi>α</mi> </semantics></math>, drone <math display="inline"><semantics> <mi>β</mi> </semantics></math>, and drone <math display="inline"><semantics> <mi>γ</mi> </semantics></math>, using the ICP algorithm, showcasing the complete coverage of the surveyed area.</p>
Full article ">Figure 6
<p>Scout v3.</p>
Full article ">Figure 7
<p>Architecture of the PSO-based tuning procedure for the SE(3) controller. The setup consists of a drone running on ROS for real-time control and state feedback, while a laptop executes the Particle Swarm Optimization (PSO) algorithm in MATLAB. Communication between the drone and the laptop enables iterative tuning of the controller parameters to optimize performance.</p>
Full article ">Figure 8
<p>Flight control optimisation process. (<b>a</b>) Real drone performing PSO-based auto tuning. (<b>b</b>) Particle Swarm Optimization (PSO) convergence graph.</p>
Full article ">Figure 9
<p>Forest site description. (<b>a</b>) A point of view of the forest site from the drone’s perspective. (<b>b</b>) Aerial view of the forest site showcasing the diverse canopy structure, ranging from dense evergreen stands to open clearings.</p>
Full article ">Figure 9 Cont.
<p>Forest site description. (<b>a</b>) A point of view of the forest site from the drone’s perspective. (<b>b</b>) Aerial view of the forest site showcasing the diverse canopy structure, ranging from dense evergreen stands to open clearings.</p>
Full article ">Figure 10
<p>Images depicting the field experiments in the forest, highlighting the multi-drone system in operation (drone <math display="inline"><semantics> <mi>α</mi> </semantics></math> in red, drone <math display="inline"><semantics> <mi>β</mi> </semantics></math> in green and drone <math display="inline"><semantics> <mi>γ</mi> </semantics></math> in blue).</p>
Full article ">Figure 11
<p>Progressive mapping of the environment by a single drone at four distinct moments during the field experiment. The figure illustrates the gradual construction of the map, depicted using a color gradient that varies with height, as the drone explores the area. Newly captured features are incrementally integrated into the overall representation.</p>
Full article ">Figure 12
<p>This figure illustrates the first inter-loop closures between two pairs of drones. These closures are crucial for ensuring cooperative mapping in multi-robot systems, reducing errors that may arise from individual robot uncertainties.</p>
Full article ">Figure 13
<p>This figure presents the frequency of inter-loop closures, revealing differences in the contributions of each drone to the overall mapping process.</p>
Full article ">Figure 14
<p>Trajectories executed by the drones during real experiments. (<b>a</b>) Variations in swarm formations over six distinct moments and overall trajectories of each individual drone. (<b>b</b>) Overlay global map (represented in red) and trajectories executed by the drones on the forest terrain.</p>
Full article ">Figure 15
<p>Maps generated by three drones, namely drone <math display="inline"><semantics> <mi>α</mi> </semantics></math>, drone <math display="inline"><semantics> <mi>β</mi> </semantics></math>, drone <math display="inline"><semantics> <mi>γ</mi> </semantics></math>, and the Global Map created by the Global Map Service. For better visualisation, a height threshold was applied and the number of points was reduced.</p>
Full article ">
24 pages, 12478 KiB  
Article
A Novel Real-Time Autonomous Localization Algorithm Based on Weighted Loosely Coupled Visual–Inertial Data of the Velocity Layer
by Cheng Liu, Tao Wang, Zhi Li and Peng Tian
Appl. Sci. 2025, 15(2), 989; https://doi.org/10.3390/app15020989 - 20 Jan 2025
Viewed by 549
Abstract
IMUs (inertial measurement units) and cameras are widely utilized and combined to autonomously measure the motion states of mobile robots. This paper presents a loosely coupled algorithm for autonomous localization, the ICEKF (IMU-aided camera extended Kalman filter), for the weighted data fusion of [...] Read more.
IMUs (inertial measurement units) and cameras are widely utilized and combined to autonomously measure the motion states of mobile robots. This paper presents a loosely coupled algorithm for autonomous localization, the ICEKF (IMU-aided camera extended Kalman filter), for the weighted data fusion of the IMU and visual measurement. The algorithm fuses motion information on the velocity layer, thereby mitigating the excessive accumulation of IMU errors caused by direct subtraction on the positional layer after quadratic integration. Furthermore, by incorporating a weighting mechanism, the algorithm allows for a flexible adjustment of the emphasis placed on IMU data versus visual information, which augments the robustness and adaptability of autonomous motion estimation for robots. The simulation and dataset experiments demonstrate that the ICEKF can provide reliable estimates for robot motion trajectories. Full article
(This article belongs to the Section Robotics and Automation)
Show Figures

Figure 1

Figure 1
<p>Relationships of the coordinate frames in the ICEKF.</p>
Full article ">Figure 2
<p>Relationships among the variables in the ICEKF state vector.</p>
Full article ">Figure 3
<p>The data flow of the variables in the ICEKF.</p>
Full article ">Figure 4
<p>Three-dimensional position curves of the simulation.</p>
Full article ">Figure 5
<p>Position error of the simulation.</p>
Full article ">Figure 6
<p>Orientation error of the simulation.</p>
Full article ">Figure 7
<p>Visual scale of the simulation.</p>
Full article ">Figure 8
<p>The initial estimate set and the converging process.</p>
Full article ">Figure 9
<p>Dataset experiments (ROOM01): (<b>a</b>) monocular ORB-SLAM, (<b>b</b>) monocular ORB-SLAM with IMU, and (<b>c</b>) monocular MSCKF.</p>
Full article ">Figure 10
<p>Dataset experiments (MainHall02): (<b>a</b>) monocular ORB-SLAM, (<b>b</b>) monocular ORB-SLAM with IMU, and (<b>c</b>) monocular MSCKF.</p>
Full article ">Figure 11
<p>Position output of the ICEKF and the ground truth (ROOM01).</p>
Full article ">Figure 12
<p>Position output of the ICEKF and the ground truth (MainHall02).</p>
Full article ">Figure 13
<p>Position error between the ICEKF and the ground truth (ROOM01).</p>
Full article ">Figure 14
<p>Orientation error between the ICEKF and the ground truth (ROOM01).</p>
Full article ">Figure 15
<p>Position error between the ICEKF and the ground truth (MainHall02).</p>
Full article ">Figure 16
<p>Orientation error between the ICEKF and the ground truth (MainHall02).</p>
Full article ">Figure 17
<p>Comparison of the partial trajectory from the ICEKF against the visual measurement with leap noise.</p>
Full article ">
20 pages, 7483 KiB  
Article
An Enhanced LiDAR-Based SLAM Framework: Improving NDT Odometry with Efficient Feature Extraction and Loop Closure Detection
by Yan Ren, Zhendong Shen, Wanquan Liu and Xinyu Chen
Processes 2025, 13(1), 272; https://doi.org/10.3390/pr13010272 - 19 Jan 2025
Viewed by 946
Abstract
Simultaneous localization and mapping (SLAM) is crucial for autonomous driving, drone navigation, and robot localization, relying on efficient point cloud registration and loop closure detection. Traditional Normal Distributions Transform (NDT) odometry frameworks provide robust solutions but struggle with real-time performance due to the [...] Read more.
Simultaneous localization and mapping (SLAM) is crucial for autonomous driving, drone navigation, and robot localization, relying on efficient point cloud registration and loop closure detection. Traditional Normal Distributions Transform (NDT) odometry frameworks provide robust solutions but struggle with real-time performance due to the high computational complexity of processing large-scale point clouds. This paper introduces an improved NDT-based LiDAR odometry framework to address these challenges. The proposed method enhances computational efficiency and registration accuracy by introducing a unified feature point cloud framework that integrates planar and edge features, enabling more accurate and efficient inter-frame matching. To further improve loop closure detection, a parallel hybrid approach combining Radius Search and Scan Context is developed, which significantly enhances robustness and accuracy. Additionally, feature-based point cloud registration is seamlessly integrated with full cloud mapping in global optimization, ensuring high-precision pose estimation and detailed environmental reconstruction. Experiments on both public datasets and real-world environments validate the effectiveness of the proposed framework. Compared with traditional NDT, our method achieves trajectory estimation accuracy increases of 35.59% and over 35%, respectively, with and without loop detection. The average registration time is reduced by 66.7%, memory usage is decreased by 23.16%, and CPU usage drops by 19.25%. These results surpass those of existing SLAM systems, such as LOAM. The proposed method demonstrates superior robustness, enabling reliable pose estimation and map construction in dynamic, complex settings. Full article
(This article belongs to the Section Manufacturing Processes and Systems)
Show Figures

Figure 1

Figure 1
<p>The system structure.</p>
Full article ">Figure 2
<p>Combined feature point cloud. (<b>a</b>) is the raw point cloud acquired by LiDAR, and (<b>b</b>) is the feature point cloud. The feature point cloud is composed of planar points, edge points, and ground points; the outlier points and small-scale points in the environment are removed; and only large-scale point clouds are retained. Compared to the original point cloud, the feature point cloud significantly reduces the number of points while effectively preserving environmental features.</p>
Full article ">Figure 3
<p>(<b>a</b>) KITTI data acquisition platform, equipped with an inertial navigation system (GPS/IMU) OXTS RT 3003, a Velodyne HDL-64E LiDAR, two 1.4 MP grayscale cameras, two 1.4 MP color cameras, and four zoom lenses. (<b>b</b>) Sensor installation positions on the platform.</p>
Full article ">Figure 4
<p>Comparison of trajectories across different algorithm frameworks for Sequence 00-10. The trajectories generated during mapping for LOAM, LeGO-LOAM, DLO, the original NDT, and our method are compared.</p>
Full article ">Figure 5
<p>Loop closure detection results for various methods on Sequence 09. It can be seen that our improved method effectively identifies the loop closure. The parallel strategy using two loop closure detection methods greatly improves detection accuracy.</p>
Full article ">Figure 6
<p>(<b>a</b>–<b>c</b>) Inter-frame registration time, memory usage, and CPU usage before and after the improvement. Our improved method effectively reduces matching time and computational load.</p>
Full article ">Figure 7
<p>Mobile robot platform.</p>
Full article ">Figure 8
<p>Maps generated using the improved method. (<b>a</b>–<b>d</b>) The one-way corridor, round-trip corridor, loop corridor, and long, feature-sparse corridor, respectively.</p>
Full article ">Figure 9
<p>(<b>a</b>–<b>d</b>) Maps generated by the original method. Significant mapping errors occurred in larger environments, such as (<b>c</b>,<b>d</b>).</p>
Full article ">Figure 9 Cont.
<p>(<b>a</b>–<b>d</b>) Maps generated by the original method. Significant mapping errors occurred in larger environments, such as (<b>c</b>,<b>d</b>).</p>
Full article ">Figure 10
<p>Detailed comparison between the improved and original methods. (<b>a</b>,<b>b</b>) The improved and original methods, respectively. The improved method balances detail preservation and computation speed, while the original sacrifices some environmental accuracy for mapping results.</p>
Full article ">Figure 11
<p>Map comparison. (<b>a</b>) The Google Earth image. (<b>b</b>) LeGO-LOAM failed to close the loop due to the lack of IMU data, leading to Z-axis drift. (<b>c</b>) The original NDT framework experienced significant drift in large-scale complex environments. (<b>d</b>) The improved method produced maps closely matching the real environment.</p>
Full article ">Figure 12
<p>Detail of Scenario 2. The improved method preserved environmental details without artifacts or mismatches.</p>
Full article ">Figure 13
<p>(<b>a</b>–<b>c</b>) Scenario 2 map comparison. (<b>b</b>) The map generated by the original NDT method lacked details. (<b>c</b>) The improved method effectively preserved details.</p>
Full article ">
8 pages, 7391 KiB  
Proceeding Paper
Comparative Analysis of LiDAR Inertial Odometry Algorithms in Blueberry Crops
by Ricardo Huaman, Clayder Gonzalez and Sixto Prado
Eng. Proc. 2025, 83(1), 9; https://doi.org/10.3390/engproc2025083009 - 9 Jan 2025
Viewed by 437
Abstract
In recent years, LiDAR Odometry (LO) and LiDAR Inertial Odometry (LIO) algorithms for robot localization have considerably improved, with significant advancements demonstrated in various benchmarks. However, their performance in agricultural environments remains underexplored. This study addresses this gap by evaluating five state-of-the-art LO [...] Read more.
In recent years, LiDAR Odometry (LO) and LiDAR Inertial Odometry (LIO) algorithms for robot localization have considerably improved, with significant advancements demonstrated in various benchmarks. However, their performance in agricultural environments remains underexplored. This study addresses this gap by evaluating five state-of-the-art LO and LIO algorithms—LeGO-LOAM, DLO, DLIO, FAST-LIO2, and Point-LIO—in a blueberry farm setting. Using an Ouster OS1-32 LiDAR mounted on a four-wheeled mobile robot, the algorithms were evaluated using the translational error metric across four distinct sequences. DLIO showed the highest accuracy across all sequences, with a minimal error of 0.126 m over a 230 m path, while FAST-LIO2 achieved its lowest translational error of 0.606 m on a U-shaped path. LeGO-LOAM, however, struggled due to the environment’s lack of linear and planar features. The results underscore the effectiveness and potential limitations of these algorithms in agricultural environments, offering insights into future improvements and adaptations. Full article
Show Figures

Figure 1

Figure 1
<p>Designated paths for each sequence followed by the robot at the blueberry farm, where each letter represents a waypoint along the trajectories.</p>
Full article ">Figure 2
<p>Wheeled mobile robot at the blueberry farm. (<b>a</b>) The robot in its initial position within an inter-row space. (<b>b</b>) The robot transitioning between blocks of crops and the separation between these marked with a yellow measuring tape.</p>
Full article ">Figure 3
<p>Estimated trajectories by each algorithm during sequences AB and AC.</p>
Full article ">Figure 4
<p>Estimated trajectories by each algorithm during sequences AD and AF.</p>
Full article ">Figure 5
<p>Close-up views of the 3D maps generated using DLIO, with each label designating the corresponding sequence from the Blueberry Crop Dataset. The path taken to create these maps is shown in yellow. The point cloud color indicates the intensity of the point return.</p>
Full article ">
17 pages, 4607 KiB  
Article
Event-Based Visual/Inertial Odometry for UAV Indoor Navigation
by Ahmed Elamin, Ahmed El-Rabbany and Sunil Jacob
Sensors 2025, 25(1), 61; https://doi.org/10.3390/s25010061 - 25 Dec 2024
Cited by 1 | Viewed by 934
Abstract
Indoor navigation is becoming increasingly essential for multiple applications. It is complex and challenging due to dynamic scenes, limited space, and, more importantly, the unavailability of global navigation satellite system (GNSS) signals. Recently, new sensors have emerged, namely event cameras, which show great [...] Read more.
Indoor navigation is becoming increasingly essential for multiple applications. It is complex and challenging due to dynamic scenes, limited space, and, more importantly, the unavailability of global navigation satellite system (GNSS) signals. Recently, new sensors have emerged, namely event cameras, which show great potential for indoor navigation due to their high dynamic range and low latency. In this study, an event-based visual–inertial odometry approach is proposed, emphasizing adaptive event accumulation and selective keyframe updates to reduce computational overhead. The proposed approach fuses events, standard frames, and inertial measurements for precise indoor navigation. Features are detected and tracked on the standard images. The events are accumulated into frames and used to track the features between the standard frames. Subsequently, the IMU measurements and the feature tracks are fused to continuously estimate the sensor states. The proposed approach is evaluated using both simulated and real-world datasets. Compared with the state-of-the-art U-SLAM algorithm, our approach achieves a substantial reduction in the mean positional error and RMSE in simulated environments, showing up to 50% and 47% reductions along the x- and y-axes, respectively. The approach achieves 5–10 ms latency per event batch and 10–20 ms for frame updates, demonstrating real-time performance on resource-constrained platforms. These results underscore the potential of our approach as a robust solution for real-world UAV indoor navigation scenarios. Full article
(This article belongs to the Special Issue Multi-sensor Integration for Navigation and Environmental Sensing)
Show Figures

Figure 1

Figure 1
<p>Workflow of the proposed event-based VIO.</p>
Full article ">Figure 2
<p>DAVIS346 event camera.</p>
Full article ">Figure 3
<p>Study area camera calibration: (<b>a</b>) a 6 × 9 chessboard with a square size of 30 mm; and (<b>b</b>) an example of the detected pattern image.</p>
Full article ">Figure 4
<p>An office environment simulation layout.</p>
Full article ">Figure 5
<p>Simulated dataset: comparison of trajectories.</p>
Full article ">Figure 6
<p>Ground-based dataset: an example of feature detection and tracking on an events accumulated frame.</p>
Full article ">Figure 7
<p>Ground-based dataset: an example of feature detection and tracking on a standard frame.</p>
Full article ">Figure 8
<p>Ground-based dataset: an example of feature detection and tracking on combined events and standard frames.</p>
Full article ">Figure 9
<p>Ground-based dataset: comparison of trajectories.</p>
Full article ">Figure 10
<p>UAV used for the experiments. (1) DAVIS346 event camera. (2) NVIDIA Jetson Xavier computer. (3) Pixhawk 4 flight controller.</p>
Full article ">Figure 11
<p>UAV-based dataset: an example of feature detection and tracking on an events accumulated frame.</p>
Full article ">Figure 12
<p>UAV-based dataset: an example of feature detection and tracking on a standard frame.</p>
Full article ">Figure 13
<p>UAV-based dataset: an example of feature detection and tracking on combined events and standard frames.</p>
Full article ">Figure 14
<p>UAV-based dataset: comparison of trajectories.</p>
Full article ">
24 pages, 31029 KiB  
Article
InCrowd-VI: A Realistic Visual–Inertial Dataset for Evaluating Simultaneous Localization and Mapping in Indoor Pedestrian-Rich Spaces for Human Navigation
by Marziyeh Bamdad, Hans-Peter Hutter and Alireza Darvishy
Sensors 2024, 24(24), 8164; https://doi.org/10.3390/s24248164 - 21 Dec 2024
Viewed by 789
Abstract
Simultaneous localization and mapping (SLAM) techniques can be used to navigate the visually impaired, but the development of robust SLAM solutions for crowded spaces is limited by the lack of realistic datasets. To address this, we introduce InCrowd-VI, a novel visual–inertial dataset specifically [...] Read more.
Simultaneous localization and mapping (SLAM) techniques can be used to navigate the visually impaired, but the development of robust SLAM solutions for crowded spaces is limited by the lack of realistic datasets. To address this, we introduce InCrowd-VI, a novel visual–inertial dataset specifically designed for human navigation in indoor pedestrian-rich environments. Recorded using Meta Aria Project glasses, it captures realistic scenarios without environmental control. InCrowd-VI features 58 sequences totaling a 5 km trajectory length and 1.5 h of recording time, including RGB, stereo images, and IMU measurements. The dataset captures important challenges such as pedestrian occlusions, varying crowd densities, complex layouts, and lighting changes. Ground-truth trajectories, accurate to approximately 2 cm, are provided in the dataset, originating from the Meta Aria project machine perception SLAM service. In addition, a semi-dense 3D point cloud of scenes is provided for each sequence. The evaluation of state-of-the-art visual odometry (VO) and SLAM algorithms on InCrowd-VI revealed severe performance limitations in these realistic scenarios. Under challenging conditions, systems exceeded the required localization accuracy of 0.5 m and the 1% drift threshold, with classical methods showing drift up to 5–10%. While deep learning-based approaches maintained high pose estimation coverage (>90%), they failed to achieve real-time processing speeds necessary for walking pace navigation. These results demonstrate the need and value of a new dataset to advance SLAM research for visually impaired navigation in complex indoor environments. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

Figure 1
<p>Sample of manual measurement process for ground-truth validation. <b>Left</b>: Real-world scene with a landmark floor tile highlighted by pink rectangle. <b>Middle</b>: Full 3D point cloud map of the scene with four adjacent floor tiles marked in blue. <b>Right</b>: Zoomed view of the marked corner of the tiles in the point cloud used for measurement.</p>
Full article ">Figure 2
<p>Correlation between real-world measurements and point-cloud-derived distances in challenging sequences, where state-of-the-art SLAM systems exhibited failure or suboptimal performance. The scatter plot demonstrates a strong linear relationship between real-world and measured distances (in centimeters), with an average error of 2.14 cm, standard deviation of 1.46 cm, and median error of 2.0 cm.</p>
Full article ">Figure 3
<p>Refined 3D reconstruction demonstrating the removal of dynamic pedestrians that initially appeared static relative to the camera on the escalator.</p>
Full article ">Figure 4
<p>Example of image data and corresponding 3D map from a dataset sequence: The top-left image shows the RGB frame, and the top-middle and top-right images represent the left and right images of a stereo pair. The bottom image shows the 3D map of the scene.</p>
Full article ">Figure 5
<p>Distribution of challenges across sequences in the InCrowd-VI dataset, categorized by crowd density levels (High: &gt;10 pedestrians per frame; Medium: 4–10 pedestrians; Low: 1–3 pedestrians; None: no pedestrians). The x-axis represents the different types of challenges, and the y-axis indicates the total number of sequences. Note that the sequences may contain multiple challenges simultaneously.</p>
Full article ">Figure 6
<p>Histogram of trajectory length.</p>
Full article ">Figure 7
<p>Example scenes from the InCrowd-VI dataset demonstrating various challenges: (<b>a</b>) high pedestrian density, (<b>b</b>) varying lighting conditions, (<b>c</b>) texture-poor surfaces, (<b>d</b>) reflective surfaces, (<b>e</b>) narrow aisles, and (<b>f</b>) stairs.</p>
Full article ">Figure 8
<p>ATE comparison of evaluated SLAM systems under challenging conditions, with the x-axis depicting sequences categorized by crowd density: high, medium, low, and none.</p>
Full article ">
26 pages, 6416 KiB  
Article
Advanced Monocular Outdoor Pose Estimation in Autonomous Systems: Leveraging Optical Flow, Depth Estimation, and Semantic Segmentation with Dynamic Object Removal
by Alireza Ghasemieh and Rasha Kashef
Sensors 2024, 24(24), 8040; https://doi.org/10.3390/s24248040 - 17 Dec 2024
Viewed by 781
Abstract
Autonomous technologies have revolutionized transportation, military operations, and space exploration, necessitating precise localization in environments where traditional GPS-based systems are unreliable or unavailable. While widespread for outdoor localization, GPS systems face limitations in obstructed environments such as dense urban areas, forests, and indoor [...] Read more.
Autonomous technologies have revolutionized transportation, military operations, and space exploration, necessitating precise localization in environments where traditional GPS-based systems are unreliable or unavailable. While widespread for outdoor localization, GPS systems face limitations in obstructed environments such as dense urban areas, forests, and indoor spaces. Moreover, GPS reliance introduces vulnerabilities to signal disruptions, which can lead to significant operational failures. Hence, developing alternative localization techniques that do not depend on external signals is essential, showing a critical need for robust, GPS-independent localization solutions adaptable to different applications, ranging from Earth-based autonomous vehicles to robotic missions on Mars. This paper addresses these challenges using Visual odometry (VO) to estimate a camera’s pose by analyzing captured image sequences in GPS-denied areas tailored for autonomous vehicles (AVs), where safety and real-time decision-making are paramount. Extensive research has been dedicated to pose estimation using LiDAR or stereo cameras, which, despite their accuracy, are constrained by weight, cost, and complexity. In contrast, monocular vision is practical and cost-effective, making it a popular choice for drones, cars, and autonomous vehicles. However, robust and reliable monocular pose estimation models remain underexplored. This research aims to fill this gap by developing a novel adaptive framework for outdoor pose estimation and safe navigation using enhanced visual odometry systems with monocular cameras, especially for applications where deploying additional sensors is not feasible due to cost or physical constraints. This framework is designed to be adaptable across different vehicles and platforms, ensuring accurate and reliable pose estimation. We integrate advanced control theory to provide safety guarantees for motion control, ensuring that the AV can react safely to the imminent hazards and unknown trajectories of nearby traffic agents. The focus is on creating an AI-driven model(s) that meets the performance standards of multi-sensor systems while leveraging the inherent advantages of monocular vision. This research uses state-of-the-art machine learning techniques to advance visual odometry’s technical capabilities and ensure its adaptability across different platforms, cameras, and environments. By merging cutting-edge visual odometry techniques with robust control theory, our approach enhances both the safety and performance of AVs in complex traffic situations, directly addressing the challenge of safe and adaptive navigation. Experimental results on the KITTI odometry dataset demonstrate a significant improvement in pose estimation accuracy, offering a cost-effective and robust solution for real-world applications. Full article
(This article belongs to the Special Issue Sensors for Object Detection, Pose Estimation, and 3D Reconstruction)
Show Figures

Figure 1

Figure 1
<p>Proposed Pipeline Architecture.</p>
Full article ">Figure 2
<p>Optical flow processed output sample for one sequence of frames.</p>
Full article ">Figure 3
<p>Sample output of depth estimation.</p>
Full article ">Figure 4
<p>Sample output of the semantic segmentation.</p>
Full article ">Figure 5
<p>Sample output of dynamic object and sky removal.</p>
Full article ">Figure 6
<p>Step-by-step preprocessing samples.</p>
Full article ">Figure 7
<p>Pose estimator architecture. I changed it and replaced the image.</p>
Full article ">Figure 8
<p>Train/Loss chart for the KITTI odometry dataset.</p>
Full article ">Figure 9
<p>The validation/Loss chart for the KITTI odometry dataset shows that the pose estimator can learn more rapidly by providing extra scene information, especially semantic segmentation, to add correction weight to each class of objects and remove dynamic ones from the estimations.</p>
Full article ">Figure 10
<p>Train and validation loss for different preprocessing stages, including no preprocessing, OF, OF with depth estimation, and OF with depth and semantic segmentation.</p>
Full article ">Figure 11
<p>Proposed model’s tracking experience output for the KITTI odometry dataset. The <span class="html-italic">X</span> and <span class="html-italic">Y</span>-axis units are in meters.</p>
Full article ">Figure 12
<p>Train/Loss with different learning rates.</p>
Full article ">Figure 13
<p>Validation/Loss with different learning rates.</p>
Full article ">
Back to TopTop