[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,608)

Search Parameters:
Keywords = mapping lidar

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
16 pages, 1905 KiB  
Article
Investigating LiDAR Metrics for Old-Growth Beech- and Spruce-Dominated Forest Identification in Central Europe
by Devara P. Adiningrat, Andrew Skidmore, Michael Schlund, Tiejun Wang, Haidi Abdullah and Marco Heurich
Remote Sens. 2025, 17(2), 251; https://doi.org/10.3390/rs17020251 - 12 Jan 2025
Viewed by 392
Abstract
Old-growth forests are essential for maintaining biodiversity, as they are formed by the complexity of diverse forest structures, such as broad variations in tree height and diameter (DBH) and conditions of living and dead trees, leading to various ecological niches. However, many efforts [...] Read more.
Old-growth forests are essential for maintaining biodiversity, as they are formed by the complexity of diverse forest structures, such as broad variations in tree height and diameter (DBH) and conditions of living and dead trees, leading to various ecological niches. However, many efforts of old-growth forest mapping from LiDAR have targeted only one specific forest structure (e.g., stand height, basal area, or stand density) by deriving information through a large number of LiDAR metrics. This study introduces a novel approach for identifying old-growth forests by optimizing a set of selected LiDAR standards and structural metrics. These metrics effectively capture the arrangement of multiple forest structures, such as canopy heterogeneity, multilayer canopy profile, and canopy openness. To determine the important LiDAR standard and structural metrics in identifying old-growth forests, multicollinearity analysis using the variance inflation factor (VIF) approach was applied to identify and remove metrics with high collinearity, followed by the random forest algorithm to rank which LiDAR standard and structural metrics are important in old-growth forest classification. The results demonstrate that the LiDAR structural metrics (i.e., advanced LiDAR metrics related to multiple canopy structures) are more important and effective in distinguishing old- and second-growth forests than LiDAR standard metrics (i.e., height- and density-based LiDAR metrics) using the European definition of a 150-year stand age threshold for old-growth forests. These structural metrics were then used as predictors for the final classification of old-growth forests, yielding an overall accuracy of 78%, with a true skill statistic (TSS) of 0.58 for the test dataset. This study demonstrates that using a few structural LiDAR metrics provides more information than a high number of standard LiDAR metrics, particularly for identifying old-growth forests in mixed temperate forests. The findings can aid forest and national park managers in developing a practical and efficient old-growth forest identification and monitoring method using LiDAR. Full article
(This article belongs to the Special Issue LiDAR Remote Sensing for Forest Mapping)
Show Figures

Figure 1

Figure 1
<p>Locations of sample plots in the Bavarian Forest National Park. The image is a true-color composite (band 432) of Sentinel-2A imagery acquired on 1 August 2020.</p>
Full article ">Figure 2
<p>The frequency of the average stand age of the old-growth and second-growth plots.</p>
Full article ">Figure 3
<p>The ranking and relative importance of the selected LiDAR metrics in old-growth forest classification generated through mean decrease accuracy (MDA) analysis. The top-ranked metrics are the metrics with the highest decreases in the accuracy coefficient. The asterisk (*) indicates the top three metrics used as the input predictors in the classification. Metrics with the □ symbol are standard metrics, and metrics with the ▲ symbol are structural metrics.</p>
Full article ">Figure 4
<p>Box plots of the Rumple index (<b>a</b>), vertical complexity index (VCI) (<b>b</b>), and gap fraction (<b>c</b>) between the old- and second-growth stages. All of the values in each selected metric were derived from the canopy height model (CHM) in the sample plot, as all the selected metrics are structural diversity metrics. Each metric can significantly distinguish old-growth from second-growth forests, determined by the Wilcoxon test <span class="html-italic">p</span>-value of &lt; 0.05.</p>
Full article ">Figure 5
<p>The SHAP summary plot of three important LiDAR metrics for discriminating old- and second-growth forests. The positive axis indicates a greater probability of a feature value to identify a class. In contrast, the negative axis indicates that the feature value has a weak contribution to identifying a class. For example, in the old-growth class, the Rumple index’s high values were distributed on the positive axis, demonstrating a high probability of old-growth emergence prediction in the classification. On the contrary, the Rumple index’s low values were distributed on the positive axis in the second-growth class, indicating that low values of the Rumple index are good predictors for the second-growth class.</p>
Full article ">
27 pages, 7047 KiB  
Article
Assessing the Impacts of Selective Logging on the Forest Understory in the Amazon Using Airborne LiDAR
by Leilson Ferreira, Edilson de Souza Bias, Quétila Souza Barros, Luís Pádua, Eraldo Aparecido Trondoli Matricardi and Joaquim J. Sousa
Forests 2025, 16(1), 130; https://doi.org/10.3390/f16010130 - 12 Jan 2025
Viewed by 210
Abstract
Reduced-impact logging (RIL) has been recognized as a promising strategy for biodiversity conservation and carbon sequestration within sustainable forest management (SFM) areas. However, monitoring the forest understory—a critical area for assessing logging impacts—remains challenging due to limitations in conventional methods such as field [...] Read more.
Reduced-impact logging (RIL) has been recognized as a promising strategy for biodiversity conservation and carbon sequestration within sustainable forest management (SFM) areas. However, monitoring the forest understory—a critical area for assessing logging impacts—remains challenging due to limitations in conventional methods such as field inventories and global navigation satellite system (GNSS) surveys, which are time-consuming, costly, and often lack accuracy in complex environments. Additionally, aerial and satellite imagery frequently underestimate the full extent of disturbances as the forest canopy obscures understory impacts. This study examines the effectiveness of the relative density model (RDM), derived from airborne LiDAR data, for mapping and monitoring understory disturbances. A field-based validation of LiDAR-derived RDM was conducted across 25 sites, totaling 5504.5 hectares within the Jamari National Forest, Rondônia, Brazil. The results indicate that the RDM accurately delineates disturbances caused by logging infrastructure, with over 90% agreement with GNSS field data. However, the model showed the greatest discrepancy for skid trails, which, despite their lower accuracy in modeling, accounted for the largest proportion of the total impacted area among infrastructure. The findings include the mapping of 35.1 km of primary roads, 117.4 km of secondary roads, 595.6 km of skid trails, and 323 log landings, with skid trails comprising the largest proportion of area occupied by logging infrastructure. It is recommended that airborne LiDAR assessments be conducted up to two years post-logging, as impacts become less detectable over time. This study highlights LiDAR data as a reliable alternative to traditional monitoring approaches, with the ability to detect understory impacts more comprehensively for monitoring selective logging in SFM areas of the Amazon, providing a valuable tool for both conservation and climate mitigation efforts. Full article
(This article belongs to the Special Issue Sustainable Management of Forest Stands)
Show Figures

Figure 1

Figure 1
<p>Location of the state of Rondônia (<b>a</b>), location of the Jamari National Forest within the state (<b>b</b>) and within the municipalities (<b>c</b>) along with the forest management units, and the selectively logged areas where LiDAR data were collected (<b>d</b>).</p>
Full article ">Figure 2
<p>A schematic representation of the relative density model (RDM) calculation showing the method for determining relative vegetation density across the understory layer (adapted from D’Oliveira et al. [<a href="#B15-forests-16-00130" class="html-bibr">15</a>]).</p>
Full article ">Figure 3
<p>LiDAR data used to map the logging impact on the forest understory: (<b>a</b>) relative density model (RDM), highlighting the changed understory in black; (<b>b</b>) each created logging infrastructure digitized by RDM and the impact zone (<b>c</b>) determined the impacted area (density 0–20). Area 12 RDM with a 1 m resolution, based on LIDAR data collected 4 months after harvest. Logging intensity of 12.8 m<sup>3</sup>.ha<sup>−1</sup> with a total impact on the understory of 14.3%.</p>
Full article ">Figure 4
<p>Digitization validation, based on LiDAR data of impact on the remaining forest obtained via mapping using GNSS measurements: (<b>a</b>) central points over the gaps of logged trees with a buffer of 5, 10, and 20 m based on field data; (<b>b</b>) skid trails superimposed on buffers of 5, 10, and 20 m based on field data; (<b>c</b>) primary and secondary roads superimposed on buffers of 5, 10, and 20 m based on field data; and (<b>d</b>) a polygon of the log landings superimposed on buffers of 5, 10, and 20 m from the central coordinate of the landing obtained in the field. The data correspond to Area 19 of the Jamari National Forest.</p>
Full article ">Figure 5
<p>Scatter plot matrix between logging-related parameters obtained with LiDAR data. The correlation coefficients in the upper triangular panel indicate the degree of correlation, with diagonal frequency histograms to visualize the data distribution in each variable. The lower triangular panel shows the scatter plots for each pair of variables in the correlation matrix, with a straight line indicating the correlation direction.</p>
Full article ">
21 pages, 3819 KiB  
Article
Improving Forest Canopy Height Mapping in Wuyishan National Park Through Calibration of ZiYuan-3 Stereo Imagery Using Limited Unmanned Aerial Vehicle LiDAR Data
by Kai Jian, Dengsheng Lu, Yagang Lu and Guiying Li
Forests 2025, 16(1), 125; https://doi.org/10.3390/f16010125 - 11 Jan 2025
Viewed by 257
Abstract
Forest canopy height (FCH) is a critical parameter for forest management and ecosystem modeling, but there is a lack of accurate FCH distribution in large areas. To address this issue, this study selected Wuyishan National Park in China as a case study to [...] Read more.
Forest canopy height (FCH) is a critical parameter for forest management and ecosystem modeling, but there is a lack of accurate FCH distribution in large areas. To address this issue, this study selected Wuyishan National Park in China as a case study to explore the calibration method for mapping FCH in a complex subtropical mountainous region based on ZiYuan-3 (ZY3) stereo imagery and limited Unmanned Aerial Vehicle (UAV) LiDAR data. Pearson’s correlation analysis, Categorical Boosting (CatBoost) feature importance analysis, and causal effect analysis were used to examine major factors causing extraction errors of digital surface model (DSM) data from ZY3 stereo imagery. Different machine learning algorithms were compared and used to calibrate the DSM and FCH results. The results indicate that the DSM extraction accuracy based on ZY3 stereo imagery is primarily influenced by slope aspect, elevation, and vegetation characteristics. These influences were particularly notable in areas with a complex topography and dense vegetation coverage. A Bayesian-optimized CatBoost model with directly calibrating the original FCH (the difference between the DSM from ZY3 and high-precision digital elevation model (DEM) data) demonstrated the best prediction performance. This model produced the FCH map at a 4 m spatial resolution, the root mean square error (RMSE) was reduced from 6.47 m based on initial stereo imagery to 3.99 m after calibration, and the relative RMSE (rRMSE) was reduced from 36.52% to 22.53%. The study demonstrates the feasibility of using ZY3 imagery for regional forest canopy height mapping and confirms the superior performance of using the CatBoost algorithm in enhancing FCH calibration accuracy. These findings provide valuable insights into the multidimensional impacts of key environmental factors on FCH extraction, supporting precise forest monitoring and carbon stock assessment in complex terrains in subtropical regions. Full article
(This article belongs to the Special Issue Mapping and Modeling Forests Using Geospatial Technologies)
Show Figures

Figure 1

Figure 1
<p>The physical features of Wuyishan National Park ((<b>a</b>)—the location with the extent of ZiYuan-3 data; (<b>b</b>)—the forest distribution in 2022 and the typical sites of UAV-LiDAR).</p>
Full article ">Figure 2
<p>Framework for mapping forest canopy height using ZiYuan-3 and UAV LiDAR data. DSM<sub>r</sub> and FCH<sub>r</sub>—digital surface model and forest canopy height from UAV LiDAR data, respectively, as reference data; DSM—the original DSM from ZiYuan-3 stereo data, FCH—original FCH (DSM-DEM).</p>
Full article ">Figure 3
<p>Relationships between investigated factors and DSM<sub>error</sub> ((<b>a</b>)—Pearson’s correlation coefficients; (<b>b</b>)—Variable importance rankings from CatBoost; (<b>c</b>)—The causal effect rankings from the causal inference analysis; DSM<sub>error</sub>—DSM residuals; LS—Slope length and steepness factor; VCI—Vegetation coverage index, EVI—Enhanced vegetation index; kNDVI—The kernel Normalized Difference Vegetation Index).</p>
Full article ">Figure 4
<p>Causal pathway diagram of various factors affecting DSM<sub>error</sub> (DSM residuals) (DSM—Digital surface model; LS—Slope length and steepness factor; VCI—Vegetation coverage index; EVI—Enhanced vegetation index; kNDVI—The kernel Normalized Difference Vegetation Index). In this diagram, the edge values quantify the causal effects: their absolute values indicate effect strength, while their signs (positive or negative) denote the direction of the relationship.</p>
Full article ">Figure 5
<p>Scatter plots for comparison of the original FCH and the calibrated FCH against FCH<sub>r</sub> across different forest types (FCH represents forest canopy height obtained by subtracting the DEM from the ZY3-derived DSM, the calibrated FCH (FCH<sub>dc</sub>) represents the canopy height obtained by the direct calibration method using CatBoost, FCH<sub>r</sub> represents the reference FCH derived from UAV LiDAR).</p>
Full article ">Figure 6
<p>Spatial distribution of forest canopy height (FCH) within Wuyishan National Park. ((<b>a</b>): FCH generated from the difference between the digital surface model from ZY3 stereo imagery and the digital elevation model from airborne LiDAR, (<b>b</b>): the calibrated FCH using the CatBoost in the direct calibration method).</p>
Full article ">Figure 7
<p>Accuracy comparison between the calibrated FCH and the ETH Global Sentinel−2 10 m Canopy Height Product (2020) [<a href="#B53-forests-16-00125" class="html-bibr">53</a>] ((<b>a</b>) FCH generated by this research, (<b>b</b>) ETH_FCH product produced by [<a href="#B53-forests-16-00125" class="html-bibr">53</a>]).</p>
Full article ">
25 pages, 13628 KiB  
Article
Gradient Enhancement Techniques and Motion Consistency Constraints for Moving Object Segmentation in 3D LiDAR Point Clouds
by Fangzhou Tang, Bocheng Zhu and Junren Sun
Remote Sens. 2025, 17(2), 195; https://doi.org/10.3390/rs17020195 - 8 Jan 2025
Viewed by 332
Abstract
The ability to segment moving objects from three-dimensional (3D) LiDAR scans is critical to advancing autonomous driving technology, facilitating core tasks like localization, collision avoidance, and path planning. In this paper, we introduce a novel deep neural network designed to enhance the performance [...] Read more.
The ability to segment moving objects from three-dimensional (3D) LiDAR scans is critical to advancing autonomous driving technology, facilitating core tasks like localization, collision avoidance, and path planning. In this paper, we introduce a novel deep neural network designed to enhance the performance of 3D LiDAR point cloud moving object segmentation (MOS) through the integration of image gradient information and the principle of motion consistency. Our method processes sequential range images, employing depth pixel difference convolution (DPDC) to improve the efficacy of dilated convolutions, thus boosting spatial information extraction from range images. Additionally, we incorporate Bayesian filtering to impose posterior constraints on predictions, enhancing the accuracy of motion segmentation. To handle the issue of uneven object scales in range images, we develop a novel edge-aware loss function and use a progressive training strategy to further boost performance. Our method is validated on the SemanticKITTI-based LiDAR MOS benchmark, where it significantly outperforms current state-of-the-art (SOTA) methods, all while working directly on two-dimensional (2D) range images without requiring mapping. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Visualization of 2D image.</p>
Full article ">Figure 2
<p>Details of improved convolution methods.</p>
Full article ">Figure 3
<p>An overview of our method. The upper part illustrates the overall workflow of the network, while the lower part details the specific implementation of each submodule.</p>
Full article ">Figure 4
<p>Details of Depth Pixel Difference Convolution (DPDC).</p>
Full article ">Figure 5
<p>Qualitative comparisons of various methods for LiDAR-MOS in different scenes on the SemanticKITTI-MOS validation set are presented. Blue circles emphasize mispredictions and indistinct boundaries. For optimal viewing, refer to the images in color and zoom in for finer details.</p>
Full article ">Figure 6
<p>Qualitative comparisons of various methods for LiDAR-MOS between consecutive frames on the SemanticKITTI-MOS validation set are presented. Blue circles emphasize mispredictions and indistinct boundaries. For optimal viewing, refer to the images in color and zoom in for finer details.</p>
Full article ">
19 pages, 11895 KiB  
Article
Mapping Spatial Variability of Sugarcane Foliar Nitrogen, Phosphorus, Potassium and Chlorophyll Concentrations Using Remote Sensing
by Ericka F. Picado, Kerin F. Romero and Muditha K. Heenkenda
Geomatics 2025, 5(1), 3; https://doi.org/10.3390/geomatics5010003 - 5 Jan 2025
Viewed by 476
Abstract
Various nutrients are needed during the sugarcane growing season for plant development and productivity. However, traditional methods for assessing nutritional status are often costly and time consuming. This study aimed to determine the level of nitrogen (N), phosphorus (P), potassium (K) and chlorophyll [...] Read more.
Various nutrients are needed during the sugarcane growing season for plant development and productivity. However, traditional methods for assessing nutritional status are often costly and time consuming. This study aimed to determine the level of nitrogen (N), phosphorus (P), potassium (K) and chlorophyll of sugarcane plants using remote sensing. Remotely sensed images were obtained using a MicaSense RedEdge-P camera attached to a drone. Leaf chlorophyll content was measured in the field using an N-Tester chlorophyll meter, and leaf samples were collected and analyzed in the laboratory for N, P and K. The highest correlation between field samples and predictor variables (spectral bands, selected vegetation indices, and plant height from Light Detection and Ranging (LiDAR)), were noted.The spatial distribution of chlorophyll, N, P, and K maps achieved 60%, 75%, 96% and 50% accuracies, respectively. The spectral profiles helped to identify areas with visual differences. Spatial variability of nutrient maps confirmed that moisture presence leads to nitrogen and potassium deficiencies, excess phosphorus, and a reduction in vegetation density (93.82%) and height (2.09 m), compared to green, healthy vegetation (97.64% density and 3.11 m in height). This robust method of assessing foliar nutrients is repeatable for the same sugarcane variety at certain conditions and leads to sustainable agricultural practices in Costa Rica. Full article
Show Figures

Figure 1

Figure 1
<p>Study area map, location and sampling plot details.</p>
Full article ">Figure 2
<p>The project workflow.</p>
Full article ">Figure 3
<p>(<b>a</b>) Orthomosaic covering one of the sample plots; and (<b>b</b>) classified orthomosaic to extract sugarcane coverage.</p>
Full article ">Figure 4
<p>Normal distribution analysis. (<b>a</b>) Index plot, (<b>b</b>) boxplot, (<b>c</b>) histogram, and (<b>d</b>) qq-plot.</p>
Full article ">Figure 5
<p>Nutritional variability. (<b>A</b>) Sugarcane coverage; (<b>B</b>) chlorophyll; (<b>C</b>) nitrogen; (<b>D</b>) phosphorus; and (<b>E</b>) potassium content.</p>
Full article ">Figure 6
<p>Average spectral profile for healthy green areas and areas with light green or brownish leaves.</p>
Full article ">Figure 7
<p>Nutritional variability map. (<b>A</b>) Chlorophyll; (<b>B</b>) nitrogen; (<b>C</b>) phosphorus; and (<b>D</b>) potassium content.</p>
Full article ">
22 pages, 1613 KiB  
Article
I-PAttnGAN: An Image-Assisted Point Cloud Generation Method Based on Attention Generative Adversarial Network
by Wenwen Li, Yaxing Chen, Qianyue Fan, Meng Yang, Bin Guo and Zhiwen Yu
Remote Sens. 2025, 17(1), 153; https://doi.org/10.3390/rs17010153 - 4 Jan 2025
Viewed by 410
Abstract
The key to building a 3D point cloud map is to ensure the consistency and accuracy of point cloud data. However, the hardware limitations of LiDAR lead to a sparse and uneven distribution of point cloud data in the edge region, which brings [...] Read more.
The key to building a 3D point cloud map is to ensure the consistency and accuracy of point cloud data. However, the hardware limitations of LiDAR lead to a sparse and uneven distribution of point cloud data in the edge region, which brings many challenges to 3D map construction, such as low registration accuracy and high construction errors in the sparse regions. To solve these problems, this paper proposes the I-PAttnGAN network to generate point clouds with image-assisted approaches, which aims to improve the density and uniformity of sparse regions and enhance the representation ability of point cloud data in sparse edge regions for distant objects. I-PAttnGAN uses the normalized flow model to extract the point cloud attention weights dynamically and then integrates the point cloud weights into image features to learn the transformation relationship between the weighted image features and the point cloud distribution, so as to realize the adaptive generation of the point cloud density and resolution. Extensive experiments are conducted on ShapeNet and nuScenes datasets. The results show that I-PAttnGAN significantly improves the performance of generating high-quality, dense point clouds in low-density regions compared with existing methods: the Chamfer distance value is reduced by about 2 times, the Earth Mover’s distance value is increased by 1.3 times, and the F1 value is increased by about 1.5 times. In addition, the effectiveness of the newly added modules is verified by ablation experiments, and the experimental results show that these modules play a key role in the generation process. Overall, the proposed model shows significant advantages in terms of accuracy and efficiency, especially in generating complete spatial point clouds. Full article
Show Figures

Figure 1

Figure 1
<p>Challenges in constructing 3D point cloud maps [<a href="#B1-remotesensing-17-00153" class="html-bibr">1</a>,<a href="#B2-remotesensing-17-00153" class="html-bibr">2</a>,<a href="#B3-remotesensing-17-00153" class="html-bibr">3</a>].</p>
Full article ">Figure 2
<p>Overall architecture. <b>In the generation module, the attention weight is calculated according to the point cloud density and weighted fusion is performed with the image data. Then, the Normalized Flows (NFs) model was used to learn the reversible transformation from complex distribution to simple Gaussian distribution. Finally, a Variational AutoEncoder (VAE) was used to model the two distributions as conditional distributions. Method In the discrimination module, the inference results of the generation module are classified and compared, and the gap between the generated point cloud data and the real point cloud data is calculated</b>.</p>
Full article ">Figure 3
<p>The point cloud generator in our model generates two important components: the Kullback–Leibler divergence term <math display="inline"><semantics> <msub> <mi mathvariant="normal">D</mi> <mrow> <mi>K</mi> <mi>L</mi> </mrow> </msub> </semantics></math>, which measures the difference between the approximate and true posterior distributions, and the normalizing flow model loss <math display="inline"><semantics> <msub> <mi mathvariant="normal">L</mi> <mrow> <mi>N</mi> <mi>F</mi> <mi>s</mi> </mrow> </msub> </semantics></math>, which helps optimize the latent space transformations. The different arrows indicate the flow of data through the generator, and the red box highlights the key components of the generator’s architecture.</p>
Full article ">Figure 4
<p>The point cloud discriminator utilizes the loss function <math display="inline"><semantics> <msub> <mi mathvariant="normal">L</mi> <mi>D</mi> </msub> </semantics></math>, which plays a critical role in adversarial training by quantifying the discriminator’s success in correctly identifying real point cloud data as real and generated point cloud data as fake. The minimization of <math display="inline"><semantics> <msub> <mi mathvariant="normal">L</mi> <mi>D</mi> </msub> </semantics></math> helps the model improve the realism of the generated point clouds. The red box highlights the key part of the discriminator’s architecture.</p>
Full article ">Figure 5
<p>Visual results of our method and the benchmark method on nuScenes when comparing the generation effects of 3D point cloud data. The red box highlights the key area of interest, showing the visualization results of our model. This part illustrates the model’s performance, specifically emphasizing the key features and improvements of the generated 3D point cloud data compared to the benchmark method.</p>
Full article ">Figure 6
<p>Results of ablation experiments conducted on the nuScenes dataset. To better demonstrate the role of the attention module in generating point cloud shape features, we visualized the inference results of these ablation experiments and compared them with the baseline model. Among them, “Ours + attention” represents the complete model structure, and “Ours” represents the model without the attention structure. The circled regions highlight areas where the point cloud’s feature representation ability significantly declines after removing the attention mechanism.</p>
Full article ">
23 pages, 26242 KiB  
Article
The Application of Fast Fourier Transform Filtering to High Spatial Resolution Digital Terrain Models Derived from LiDAR Sensors for the Objective Mapping of Surface Features and Digital Terrain Model Evaluations
by Alberto González-Díez, Ignacio Díaz-Martínez, Pablo Cruz-Hernández, Antonio Barreda-Argüeso and Matthew Doughty
Remote Sens. 2025, 17(1), 150; https://doi.org/10.3390/rs17010150 - 4 Jan 2025
Viewed by 466
Abstract
In this paper, the application is investigated of fast Fourier transform filtering (FFT-FR) to high spatial resolution digital terrain models (HR-DTM) derived from LiDAR sensors, assessing its efficacy in identifying genuine relief elements, including both natural geological features and anthropogenic landforms. The suitability [...] Read more.
In this paper, the application is investigated of fast Fourier transform filtering (FFT-FR) to high spatial resolution digital terrain models (HR-DTM) derived from LiDAR sensors, assessing its efficacy in identifying genuine relief elements, including both natural geological features and anthropogenic landforms. The suitability of the derived filtered geomorphic references (FGRs) is evaluated through spatial correlation with ground truths (GTs) extracted from the topographical and geological geodatabases of Santander Bay, Northern Spain. In this study, it is revealed that existing artefacts, derived from vegetation or human infrastructures, pose challenges in the units’ construction, and large physiographic units are better represented using low-pass filters, whereas detailed units are more accurately depicted with high-pass filters. The results indicate a propensity of high-frequency filters to detect anthropogenic elements within the DTM. The quality of GTs used for validation proves more critical than the geodatabase scale. Additionally, in this study, it is demonstrated that the footprint of buildings remains uneliminated, indicating that the model is a poorly refined digital surface model (DSM) rather than a true digital terrain model (DTM). Experiments validate the DTM’s capability to highlight contacts and constructions, with water detection showing high precision (≥60%) and varying precision for buildings. Large units are better captured with low filters, whilst high filters effectively detect anthropogenic elements and more detailed units. This facilitates the design of validation and correction procedures for DEMs derived from LiDAR point clouds, enhancing the potential for more accurate and objective Earth surface representation. Full article
Show Figures

Figure 1

Figure 1
<p>Area selected in the three scenarios framed within Santander Bay (Zone_1). The heights proceed from a clip of the DTM called 9400_MDT02-ETRS89-HU30-0035-1-COB2.tif (IGN_2024) elaborated from the 2nd coverture LiDAR (acquired from 2015 to present). (<b>a</b>) Hillshade of the DTM used (the model has an ETRS89 geographic coordinate system with UTM planimetric projection). Inner line purple box corresponds to the area selected to carry out the experiments included in the scenarios 1 and 2. Grey boxes in this zone correspond to the areas selected in the Scenario_3 (a and b). (<b>b</b>) Principal land-use units considered in this study (A, facilities; B, roads and infrastructures; C, buildings; D, natural slopes; and E, water) extracted from BTN [<a href="#B36-remotesensing-17-00150" class="html-bibr">36</a>,<a href="#B37-remotesensing-17-00150" class="html-bibr">37</a>].</p>
Full article ">Figure 2
<p>Ground truths (GTs) utilized in scenarios 1 and 2. (<b>a</b>) Distinction between the ocean and continent (ocean valued as 100, continent valued as 200); (<b>b</b>) delineation of buildings and urban areas (buildings valued as 100, natural terrains valued as 200, water valued as 300); (<b>c</b>) identification of bedrock units (bedrocks valued as 100, remaining units valued as 200); (<b>d</b>) identification of superficial units (superficial units valued as 100, remaining units valued as 200); (<b>e</b>) identification of anthropic landscapes (anthropic landscapes valued as 100, remaining units valued as 200); (<b>f</b>) identification of dolines and karstic depressions (dolines and depressions valued as 100, remaining units valued as 200).</p>
Full article ">Figure 3
<p>Four aspects of the elevation analysis of the Santander Bay area (Zone_1) are presented. (<b>a</b>) The DTM utilized emphasizes elevations in a color range (−24.8 to 130.5 m), the black line box corresponds to the area selected in the three scenarios (the heights in this area ranging from −1.02 to 52.87 m); (<b>b</b>) a histogram depicting the distribution of elevations present in the Santander Bay area (mean and standard deviation are indicated); (<b>c</b>) a general view of the magnitude-true frequency plot, with cut-off frequencies (COFs) accentuated (green and orange dots are maximum and minimum of the main harmonics, respectively), the COF with figures are the filtered radius (FR) considered in this study; (<b>d</b>) a detailed view of the magnitude-true frequency plot, showcasing low and medium frequencies, with the COFs emphasized by figures.</p>
Full article ">Figure 4
<p>A visual comparison of the filtered geomorphic references (FGRs) derived in the area of Santander Bay for each cut-off frequency (COF), or filter radius, (FR) is presented. Color vectors illustrate the FGR (in the legend, each FGR is identified as the frequency symbol, FC plus FR figure), all of which are displayed on a shaded relief extracted from <a href="#remotesensing-17-00150-f001" class="html-fig">Figure 1</a> and superimposed in black.</p>
Full article ">Figure 5
<p>A visual comparison is shown between the filtered geomorphic reference models (FGRMs) from Experiment_1 of Scenario_1 and the ground truth (GT) used to distinguish between continental and sea areas (<a href="#remotesensing-17-00150-f002" class="html-fig">Figure 2</a>a). The figure is divided into the following subdivisions: (<b>a</b>) At the top of the rows, FGRMs for each filtering radius, FR (COF in the text), are indicated (each box considered includes the frequency symbol, FC, plus filter figure). The units represented are (1) area detected by the filter as water, and (2) area detected as continent; at the bottom of the rows, the corresponding validation model obtained appears (the spatial intersection units are 101, 102, 201, 202, with 101 and 202 being the hits (H), while the rest are failures). The four rows located at the right side show equivalent results obtained for each FR considered. (<b>b</b>) Global accuracy, GA, and kappa values obtained from each spatial intersection, emphasizing the FR offering the best accuracy indexes.</p>
Full article ">Figure 6
<p>Visual comparison between the filtered geomorphic reference models (FGRMs) and the spatial validation models derived for the analysis of three land use classes (Experiment_2, Scenario_1). (<b>a</b>) At the top of the rows, FGRMs for each filtering radius, FR (COF in the text), are indicated (each box considered includes the frequency symbol, FC, plus filter figure). The units represented are as follows: (1) area detected by the filter as buildings; (2) area detected as natural slopes; and (3) area detected as water. At the bottom of the rows, the derived validation models appear (the spatial intersection units are: 101, 102, 103, 201, 202, 203, 301, 302, 303, with 101, 202, and 303 being the hits or H, while the rest are failures). (<b>b</b>) Global accuracy, GA, and kappa values derived from each spatial intersection, emphasizing the FR offering the best accuracy indexes.</p>
Full article ">Figure 7
<p>Validation models obtained in Experiment_3 by the spatial crossing between the GT units of substrate rocks and corresponding FGRMs presented in <a href="#app1-remotesensing-17-00150" class="html-app">Figure S1B</a>. On the left, a true magnitude–frequency plot highlighting the result that offers the best accuracy indexes (global accuracy, GA, and kappa). Positions of the frequencies used in filtering or filter radius (FR) are introduced to aid identification. On the right, validation models obtained for each spatial crossing with the FGRMs considered. Filters are presented using acronyms FC plus FR (or filter figure).</p>
Full article ">Figure 8
<p>Validation models obtained in Experiment_4 by the spatial crossing between the GT units of surface units and corresponding FGRMs presented in <a href="#remotesensing-17-00150-f002" class="html-fig">Figure 2</a>d. On the left, a magnitude–frequency plot highlighting the result that offers the best accuracy indexes (global accuracy, GA, and kappa). Positions of the frequencies used in filtering or filter radius (FR) are introduced to aid identification. Filters are presented using acronyms FC plus FR (or filter figure).</p>
Full article ">Figure 9
<p>Validation models obtained in Experiment_5 by the spatial crossing between the GT units of anthropic reliefs and corresponding FGRMs presented in <a href="#app1-remotesensing-17-00150" class="html-app">Figure S3B</a>. On the left, a magnitude-frequency plot highlighting the result that offers the best accuracy indexes (global accuracy, GA, and kappa). Positions of the frequencies used in filtering or filter radius (FR) are introduced to aid identification. Filters are presented using acronyms FC plus FR (or filter figure).</p>
Full article ">Figure 10
<p>Models obtained in Experiment_6 by the spatial crossing between the GT units of dolines and the corresponding FGRMs presented in <a href="#app1-remotesensing-17-00150" class="html-app">Figure S4B</a>. On the left, a magnitude–frequency plot highlighting the result that offers the best accuracy indexes (global accuracy, GA, and kappa). Positions of the frequencies used in filtering or filter radius (FR) are introduced to aid identification. F Filters are presented using acronyms FC plus FR (or filter figure).</p>
Full article ">Figure 11
<p>Comparison between the GT of the new surface model and its corresponding FGRMs (Experiment_7). (<b>a</b>) A new GT model obtained for the surface units in the selected study area. (<b>b</b>) Detail of the GT model showing the incorporated geomorphic contacts (black lines) corresponding to the following elements: coastline, depressions, doline fields, and isolated dolines. (<b>c</b>) Filtered geomorphic reference modes (FGRMs) obtained by applying the FFT filters, whose filter radius (FR or cut-off frequencies) appear in the header. Filters are presented using acronyms FC plus FR (or filter figure).</p>
Full article ">Figure 12
<p>Validation models obtained in Experiment_7 by the spatial crossing of the new surface units ground truth (GT) and the corresponding FGRMs presented in <a href="#remotesensing-17-00150-f011" class="html-fig">Figure 11</a>c. On the left, a magnitude–frequency plot highlighting the result that offers the best accuracy indexes (global accuracy, GA, and kappa). Positions of the frequencies used in filtering or filter radius (FR) are introduced to help their identification. Filters are presented using acronyms FC plus FR (or filter figure).</p>
Full article ">Figure 13
<p>Validation of the GT generated by buildings and constructions (Experiment_8) in an urban location (black box area B of <a href="#remotesensing-17-00150-f001" class="html-fig">Figure 1</a>). The respective FGRMs obtained are not presented in the figure. Buildings have a value of 100, the remaining classes receive 200. (<b>a</b>) GT model obtained for this classification along Zone_1, where urban units have the value of 100 and the remaining receive 200. (<b>b</b>) Detail of the GT model for the urban area B (<a href="#remotesensing-17-00150-f001" class="html-fig">Figure 1</a>, area b), using the same values as the previous. (<b>c</b>) Validation models obtained from the spatial crossing between GT and FGRMs. Units are hits (101 and 202); and failures are 102 and 201. On the right, a table with the accuracy index values obtained for the spatial crossing. Filters are presented using acronyms FC plus FR (or filter figure).</p>
Full article ">
22 pages, 14296 KiB  
Article
Calibration-Enhanced Multi-Awareness Network for Joint Classification of Hyperspectral and LiDAR Data
by Quan Zhang, Zheyuan Cui, Tianhang Wang, Zhaoxin Li and Yifan Xia
Electronics 2025, 14(1), 102; https://doi.org/10.3390/electronics14010102 - 30 Dec 2024
Viewed by 340
Abstract
Hyperspectral image (HSI) and light detection and ranging (LiDAR) data joint classification has been applied in the field of ground category recognition. However, existing methods still perform poorly in extracting high-dimensional features and elevation information, resulting in insufficient data classification accuracy. To address [...] Read more.
Hyperspectral image (HSI) and light detection and ranging (LiDAR) data joint classification has been applied in the field of ground category recognition. However, existing methods still perform poorly in extracting high-dimensional features and elevation information, resulting in insufficient data classification accuracy. To address this challenge, we propose a novel and efficient Calibration-Enhanced Multi-Awareness Network (CEMA-Net), which exploits the joint spectral–spatial–elevation features in depth to realize the accurate identification of land cover categories. Specifically, we propose a novel multi-way feature retention (MFR) module that explores deep spectral–spatial–elevation semantic information in the data through multiple paths. In addition, we propose spectral–spatial-aware enhancement (SAE) and elevation-aware enhancement (EAE) modules, which effectively enhance the awareness of ground objects that are sensitive to spectral and elevation information. Furthermore, to address the significant representation disparities and spatial misalignments between multi-source features, we propose a spectral–spatial–elevation feature calibration fusion (SFCF) module to efficiently integrate complementary characteristics from heterogeneous features. It incorporates two key advantages: (1) efficient learning of discriminative features from multi-source data, and (2) adaptive calibration of spatial differences. Comparative experimental results on the MUUFL, Trento, and Augsburg datasets demonstrate that CEMA-Net outperforms existing state-of-the-art methods, achieving superior classification accuracy with better feature map precision and minimal noise. Full article
(This article belongs to the Special Issue Advances in AI Technology for Remote Sensing Image Processing)
Show Figures

Figure 1

Figure 1
<p>Overview of the CEMA-Net Architecture.</p>
Full article ">Figure 2
<p>Illustration of spectral–spatial-aware enhancement and elevation-aware enhancement.</p>
Full article ">Figure 3
<p>Illustration of Spectral–Spatial–Elevation Calibration Fusion Module.</p>
Full article ">Figure 4
<p>Visual depiction of the MUUFL dataset: (<b>a</b>) False-color composite image of the hyperspectral data. (<b>b</b>) Ground truth map showcasing the various land cover categories. (<b>c</b>) indicates the feature category. Red box indicates the enlarged display.</p>
Full article ">Figure 5
<p>Visual depiction of the Trento dataset. (<b>a</b>) False-color composite image of the hyperspectral data. (<b>b</b>) Ground truth map showcasing the various land cover categories. (<b>c</b>) indicates the feature category. Red box indicates the enlarged display.</p>
Full article ">Figure 6
<p>Visual depiction of the Augsburg dataset. (<b>a</b>) False-color composite image of the hyperspectral data. (<b>b</b>) Ground truth map showcasing the various land cover categories. (<b>c</b>) indicates the feature category. Red box indicates the enlarged display.</p>
Full article ">Figure 7
<p>Impact of patch size on classification metrics: OA, AA, and kappa coefficient. (<b>a</b>) MUUFL. (<b>b</b>) Trento. (<b>c</b>) Augsburg.</p>
Full article ">Figure 8
<p>Impact of reduced spectral dimensionality on classification metrics: OA, AA, and kappa coefficient. (<b>a</b>) MUUFL. (<b>b</b>) Trento. (<b>c</b>) Augsburg.</p>
Full article ">Figure 9
<p>Impact of learning rate on classification metrics: OA, AA, and kappa coefficient. (<b>a</b>) MUUFL. (<b>b</b>) Trento. (<b>c</b>) Augsburg.</p>
Full article ">Figure 10
<p>Maps depicting the classification of MUUFL using various methods. (<b>a</b>) RF. (<b>b</b>) SVM. (<b>c</b>) 2D-CNN. (<b>d</b>) HybridSN. (<b>e</b>) GAHT. (<b>f</b>) CoupledCNN. (<b>g</b>) CALC. (<b>h</b>) HCTnet. (<b>i</b>) M2FNet. (<b>j</b>) our CEMA-Net. Red box indicates the enlarged display.</p>
Full article ">Figure 11
<p>Maps depicting the classification of Trento using various methods. (<b>a</b>) RF. (<b>b</b>) SVM. (<b>c</b>) 2D-CNN. (<b>d</b>) HybridSN. (<b>e</b>) GAHT. (<b>f</b>) CoupledCNN. (<b>g</b>) CALC. (<b>h</b>) HCTnet. (<b>i</b>) M2FNet. (<b>j</b>) our CEMA-Net. Red box indicates the enlarged display.</p>
Full article ">Figure 11 Cont.
<p>Maps depicting the classification of Trento using various methods. (<b>a</b>) RF. (<b>b</b>) SVM. (<b>c</b>) 2D-CNN. (<b>d</b>) HybridSN. (<b>e</b>) GAHT. (<b>f</b>) CoupledCNN. (<b>g</b>) CALC. (<b>h</b>) HCTnet. (<b>i</b>) M2FNet. (<b>j</b>) our CEMA-Net. Red box indicates the enlarged display.</p>
Full article ">Figure 12
<p>Maps depicting the classification of Augsburg using various methods. (<b>a</b>) RF. (<b>b</b>) SVM. (<b>c</b>) 2D-CNN. (<b>d</b>) HybridSN. (<b>e</b>) GAHT. (<b>f</b>) CoupledCNN. (<b>g</b>) CALC. (<b>h</b>) HCTnet. (<b>i</b>) M2FNet. (<b>j</b>) Our CEMA-Net. Red box indicates the enlarged display.</p>
Full article ">
20 pages, 5327 KiB  
Article
Using a YOLO Deep Learning Algorithm to Improve the Accuracy of 3D Object Detection by Autonomous Vehicles
by Ramavhale Murendeni, Alfred Mwanza and Ibidun Christiana Obagbuwa
World Electr. Veh. J. 2025, 16(1), 9; https://doi.org/10.3390/wevj16010009 - 27 Dec 2024
Viewed by 667
Abstract
This study presents an adaptation of the YOLOv4 deep learning algorithm for 3D object detection, addressing a critical challenge in autonomous vehicle (AV) systems: accurate real-time perception of the surrounding environment in three dimensions. Traditional 2D detection methods, while efficient, fall short in [...] Read more.
This study presents an adaptation of the YOLOv4 deep learning algorithm for 3D object detection, addressing a critical challenge in autonomous vehicle (AV) systems: accurate real-time perception of the surrounding environment in three dimensions. Traditional 2D detection methods, while efficient, fall short in providing the depth and spatial information necessary for safe navigation. This research modifies the YOLOv4 architecture to predict 3D bounding boxes, object depth, and orientation. Key contributions include introducing a multi-task loss function that optimizes 2D and 3D predictions and integrating sensor fusion techniques that combine RGB camera data with LIDAR point clouds for improved depth estimation. The adapted model, tested on real-world datasets, demonstrates a significant increase in 3D detection accuracy, achieving a mean average precision (mAP) of 85%, intersection over union (IoU) of 78%, and near real-time performance at 93–97% for detecting vehicles and 75–91% for detecting people. This approach balances high detection accuracy and real-time processing, making it highly suitable for AV applications. This study advances the field by showing how an efficient 2D detector can be extended to meet the complex demands of 3D object detection in real-world driving scenarios without sacrificing computational efficiency. Full article
(This article belongs to the Special Issue Motion Planning and Control of Autonomous Vehicles)
Show Figures

Figure 1

Figure 1
<p>YOLOv3, reprinted from Ref [<a href="#B32-wevj-16-00009" class="html-bibr">32</a>].</p>
Full article ">Figure 2
<p>SSD network structure, reprinted from Ref [<a href="#B32-wevj-16-00009" class="html-bibr">32</a>].</p>
Full article ">Figure 3
<p>RetinaNet structure, reprinted from Ref [<a href="#B32-wevj-16-00009" class="html-bibr">32</a>].</p>
Full article ">Figure 4
<p>Sample Image 1.</p>
Full article ">Figure 5
<p>In the Image, we can see that the camera-based and deep learning-based detection methods function together perfectly, as the image taken by a camera is easily detected, and the cars were detected, as indicated by the addition of boxes around them.</p>
Full article ">Figure 6
<p>The YOLO deep learning algorithm proved to be a viable option for enhancing the accuracy of 3D object identification systems in self-driving vehicles.</p>
Full article ">Figure 7
<p>Sample Image 2.</p>
Full article ">Figure 8
<p>The sample image includes cars and people; the three closest people were identified with confidence scores between 91% and 79%, which is very high and good, and the four closest vehicles were identified with confidence scores ranging from 95% to 93%.</p>
Full article ">Figure 9
<p>Sample Image 3.</p>
Full article ">
18 pages, 12334 KiB  
Article
Canopy Height Integration for Precise Forest Aboveground Biomass Estimation in Natural Secondary Forests of Northeast China Using Gaofen-7 Stereo Satellite Data
by Caixia Liu, Huabing Huang, Zhiyu Zhang, Wenyi Fan and Di Wu
Remote Sens. 2025, 17(1), 47; https://doi.org/10.3390/rs17010047 - 27 Dec 2024
Viewed by 432
Abstract
Accurate estimates of forest aboveground biomass (AGB) are necessary for the accurate tracking of forest carbon stock. Gaofen-7 (GF-7) is the first civilian sub-meter three-dimensional (3D) mapping satellite from China. It is equipped with a laser altimeter system and a dual-line array stereoscopic [...] Read more.
Accurate estimates of forest aboveground biomass (AGB) are necessary for the accurate tracking of forest carbon stock. Gaofen-7 (GF-7) is the first civilian sub-meter three-dimensional (3D) mapping satellite from China. It is equipped with a laser altimeter system and a dual-line array stereoscopic mapping camera, which enables it to synchronously generate full-waveform LiDAR data and stereoscopic images. The bulk of existing research has examined how accurate GF-7 is for topographic measurements of bare land or canopy height. The measurement of forest aboveground biomass has not received as much attention as it deserves. This study aimed to assess the GF-7 stereo imaging capability, displayed as topographic features for aboveground biomass estimation in forests. The aboveground biomass model was constructed using the random forest machine learning technique, which was accomplished by combining the use of in situ field measurements, pairs of GF-7 stereo images, and the corresponding generated canopy height model (CHM). Findings showed that the biomass estimation model had an accuracy of R2 = 0.76, RMSE = 7.94 t/ha, which was better than the inclusion of forest canopy height (R2 = 0.30, RMSE = 21.02 t/ha). These results show that GF-7 has considerable application potential in gathering large-scale high-precision forest aboveground biomass using a restricted amount of field data. Full article
Show Figures

Figure 1

Figure 1
<p>Location map of the study area (Shangzhi, Heilongjiang, China). (<b>a</b>) The location of the study area; (<b>b</b>) field plots over the GF-7 multispectral image on 20 August 2020 (R—red, G—green, B—blue).</p>
Full article ">Figure 2
<p>The procedure for calculating forest canopy height and biomass from GF-7 stereoscopic imagery.</p>
Full article ">Figure 3
<p>The August and November DSM and CHM. This figure shows only the DSM and CHM for the common regions between August and November, highlighted by the read box. (<b>a</b>,<b>b</b>) DSMs for August and November, respectively, and (<b>e</b>,<b>f</b>) show the larger detail plots in the red boxes. (<b>c</b>,<b>d</b>) CHMs for August and November, respectively, and (<b>g</b>,<b>h</b>) show the larger detail plots in the red boxes.</p>
Full article ">Figure 4
<p>The scatter plot illustrates the relationship between canopy heights predicted using a canopy height model and field-measured heights for two different time points: August and November. The light green points and corresponding regression line represent the August data, while the light blue points and their regression line represent the November data. The 1:1 line (grey dashed) indicates perfect concordance between predicted and measured heights.</p>
Full article ">Figure 5
<p>Feature importance scores for predicting AGB using random forest models under two scenarios: S1 and S2. Features are ranked in decreasing order of importance based on the mean decrease in mean squared error (MSE). The feature “class” refers to land cover classification data, distinguishing between forested and non-forested areas, derived from geographic national condition data.</p>
Full article ">Figure 6
<p>Predicted biomass maps in different scenarios: (<b>a</b>) for S1 scenario and (<b>b</b>) for S2 scenario. Detailed drawings of the red-framed area are shown in <a href="#remotesensing-17-00047-f008" class="html-fig">Figure 8</a>.</p>
Full article ">Figure 7
<p>Scatter plots depicting the relationship between predicted biomass (t/ha) and field-measured biomass (t/ha) for four different scenarios (S1, S2, S3, and S4) in 2020. Detailed scenario descriptions are provided in <a href="#remotesensing-17-00047-t002" class="html-table">Table 2</a>. (<b>a</b>) Scatter plot includes a regression line, with annotations displaying the regression equation, coefficient of determination (R<sup>2</sup>), and root mean square error (RMSE) to quantitatively assess model performance. (<b>b</b>) Residuals for each model’s prediction compared with field biomass. The results demonstrate incremental improvements in biomass prediction accuracy from S1 to S4, highlighting the significant impact of incorporating CHM and DSM data. Scenarios S2 and S3 show enhanced prediction accuracy due to the inclusion of detailed canopy height information.</p>
Full article ">Figure 8
<p>Biomass, spectrum, and canopy height spatial features at the same location. The biomass detail map on the far left shows the result of the S1 scenario, while the one on the right shows the result of the S2 scenario. The biomass ramp is consistent with that shown in <a href="#remotesensing-17-00047-f006" class="html-fig">Figure 6</a>, and the RGB channel denotes the real color channel display of GF-7. The CHM ramp is similar to those shown in <a href="#remotesensing-17-00047-f003" class="html-fig">Figure 3</a>. The last column shows land cover, with black indicating forested area and white representing non-forest land. Figure (<b>a</b>): the disturbance of water bodies and soil moisture on the river valley delta causes the forest vegetation spectra to be mistaken for bare soil and water bodies, which leads to an underestimating of biomass forecast based solely on spectral properties. The regional variability of forest species and height under various topographic circumstances is depicted in Figure (<b>b</b>). Because of the region’s eastern side’s relative flatness, low forest heights, and predominance of coniferous tree species, biomass estimations that consider CHM factors are more accurate in reflecting the real distribution. Figure (<b>c</b>): due to the overestimation of the height of the farmland vegetation caused by spectral features alone (August during the growing season), the farmland’s spectrum is viewed as being spectral like the forest. In agriculture, the average biomass is less than 5 t/ha, although biomass is more precisely anticipated because of the constraint that CHM is approximately 0 t/ha. A logging site is in the region of figure (<b>d</b>), with low biomass. A more accurate prediction of the biomass distribution is made by taking canopy height features into account.</p>
Full article ">
21 pages, 6770 KiB  
Article
Revolutionizing RIS Networks: LiDAR-Based Data-Driven Approach to Enhance RIS Beamforming
by Ahmad M. Nazar, Mohamed Y. Selim and Daji Qiao
Sensors 2025, 25(1), 75; https://doi.org/10.3390/s25010075 - 26 Dec 2024
Viewed by 407
Abstract
Reconfigurable Intelligent Surface (RIS) panels have garnered significant attention with the emergence of next-generation network technologies. This paper proposes a novel data-driven approach that leverages Light Detecting and Ranging (LiDAR) sensors to enhance user localization and beamforming in RIS-assisted networks. Integrating LiDAR sensors [...] Read more.
Reconfigurable Intelligent Surface (RIS) panels have garnered significant attention with the emergence of next-generation network technologies. This paper proposes a novel data-driven approach that leverages Light Detecting and Ranging (LiDAR) sensors to enhance user localization and beamforming in RIS-assisted networks. Integrating LiDAR sensors into the network will be instrumental, offering high-speed and precise 3D mapping capabilities, even in low light or adverse weather conditions. LiDAR data facilitate user localization, enabling the determination of optimal RIS coefficients. Our approach extends a Graph Neural Network (GNN) by integrating LiDAR-captured user locations as inputs. This extension enables the GNN to effectively learn the mapping from received pilots to optimal beamformers and reflection coefficients to maximize the RIS-assisted sumrate among multiple users. The permutation-equivariant and -invariant properties of the GNN proved advantageous in efficiently handling the LiDAR data. Our simulation results demonstrated significant improvements in sum rates compared with conventional methods. Specifically, including locations improved on excluding locations by up to 25% and outperformed the Linear Minimum Mean Squared Error (LMMSE) channel estimation by up to 85% with varying downlink power and 98% with varying pilot lengths, and showed a remarkable 190% increase with varying downlink power compared with scenarios excluding the RIS. Full article
Show Figures

Figure 1

Figure 1
<p>The general flow of the proposed system where (<b>A</b>) shows the UE pilot transmissions sent to the GNN in the uplink phase, and LiDAR captures the environment for user detection. In (<b>B</b>), the GNN optimizes the RIS phase shifts and BS beamforming vectors. In (<b>C</b>), the RIS microcontroller is reconfigured for the optimal phase coefficients, and the BS selects the optimal beamformers for the optimal UE downlink sumrate.</p>
Full article ">Figure 2
<p>Dynamic environment scenario with blockages, active users, and detection framework uplink pilot transmissions for optimized downlink sumrate. UE attempt to send uplink pilot sequences, as indicated by the orange arrow, directly to the BS but are blocked by an obstacle; the UE then use the RIS in LoS with the BS and the UE to reflect the uplink pilot sequence. The BS uses the received uplink pilot sequence to optimal beamforming vectors and RIS phase coefficients for optimal UE sumrate in the downlink as shown by the green arrow.</p>
Full article ">Figure 3
<p>Definition of angles in relation to the RIS plane, user positions, and beamforming vectors. The RIS comprises a uniform rectangular array configured as <math display="inline"><semantics> <mrow> <mi>n</mi> <mo>×</mo> <mi>n</mi> </mrow> </semantics></math> and is positioned on the <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>y</mi> <mo>,</mo> <mi>z</mi> <mo>)</mo> </mrow> </semantics></math>-plane. The direct BS-to-UE signal is blocked by an obstacle as shown by the orange arrow, and the UE-RIS-BS communication is established to optimize the UE sumrate as indicated by the green arrow.</p>
Full article ">Figure 4
<p>High-level GNN architecture mapping user pilots to beamforming vectors and phase shifts. Inputs pass through a fully connected neural network, <span class="html-italic">D</span> combination and aggregation layers, a linear layer, and a normalization layer to output phase coefficients and optimal beamforming vectors.</p>
Full article ">Figure 5
<p>Overview of SFA3D framework model and components.</p>
Full article ">Figure 6
<p>Sumrate comparison with various downlink powers <math display="inline"><semantics> <msub> <mi>P</mi> <mi>downlink</mi> </msub> </semantics></math>.</p>
Full article ">Figure 7
<p>Sumrate comparison with various uplink powers <math display="inline"><semantics> <mrow> <mo>{</mo> <msub> <mi>P</mi> <mi>uplink</mi> </msub> <mo>}</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 8
<p>Sumrate comparison with varying pilot lengths within {L}.</p>
Full article ">Figure 9
<p>Sumrate comparison with varying RIS elements within {N}.</p>
Full article ">
21 pages, 6133 KiB  
Article
BEV Semantic Map Reconstruction for Self-Driving Cars with the Multi-Head Attention Mechanism
by Yi-Cheng Liao, Jichiang Tsai and Hsuan-Ying Chien
Electronics 2025, 14(1), 32; https://doi.org/10.3390/electronics14010032 - 25 Dec 2024
Viewed by 404
Abstract
Environmental perception is crucial for safe autonomous driving, enabling accurate analysis of the vehicle’s surroundings. While 3D LiDAR is traditionally used for 3D environment reconstruction, its high cost and complexity present challenges. In contrast, camera-based cross-view frameworks can offer a cost-effective alternative. Hence, [...] Read more.
Environmental perception is crucial for safe autonomous driving, enabling accurate analysis of the vehicle’s surroundings. While 3D LiDAR is traditionally used for 3D environment reconstruction, its high cost and complexity present challenges. In contrast, camera-based cross-view frameworks can offer a cost-effective alternative. Hence, this manuscript proposes a new cross-view model to extract mapping features from camera images and then transfer them to a Bird’s-Eye View (BEV) map. Particularly, a multi-head attention mechanism in the decoder architecture generates the final semantic map. Each camera learns embedding information corresponding to its position and angle within the BEV map. Cross-view attention fuses information from different perspectives to predict top-down map features enriched with spatial information. The multi-head attention mechanism then globally performs dependency matches, enhancing long-range information and capturing latent relationships between features. Transposed convolution replaces traditional upsampling methods, avoiding high similarities of local features and facilitating semantic segmentation inference of the BEV map. Finally, we conduct numerous simulation experiments to verify the performance of our cross-view model. Full article
(This article belongs to the Special Issue Advancement on Smart Vehicles and Smart Travel)
Show Figures

Figure 1

Figure 1
<p>Overall process flow.</p>
Full article ">Figure 2
<p>MSA block architecture diagram.</p>
Full article ">Figure 3
<p>The operation of the transposed convolution.</p>
Full article ">Figure 4
<p>Transposed convolution upsampling network architecture.</p>
Full article ">Figure 5
<p>The operation of the dilated convolution.</p>
Full article ">Figure 6
<p>Dilated convolution upsampling network architecture.</p>
Full article ">Figure 7
<p>Attention skip connection layer network architecture.</p>
Full article ">Figure 8
<p>Comparisons of different algorithms: (<b>a</b>) the original input; (<b>b</b>) the result of the CVT; (<b>c</b>) the result of using MSA; and (<b>d</b>) the result of using both MSA and transposed convolution.</p>
Full article ">Figure 9
<p>The effect of DCU: (<b>a</b>) the original input; (<b>b</b>) the result without applying DCU; and (<b>c</b>) the result with applying DCU.</p>
Full article ">Figure 10
<p>The impact of incorporating attention skip connection layers: (<b>a</b>) the original input; (<b>b</b>) the result of using standard skip connection layers; and (<b>c</b>) the result of using attention skip connection layers.</p>
Full article ">Figure 11
<p>The training loss curves for both the CVT and the proposed model.</p>
Full article ">Figure 12
<p>The impact of filtering low-confidence predictions for both the CVT and the proposed model.</p>
Full article ">Figure 13
<p>A qualitative comparison of BEV map predictions with CVT in diverse environments: (<b>a</b>) daytime; (<b>b</b>) daytime; (<b>c</b>) nighttime; and (<b>d</b>) rainy weather.</p>
Full article ">Figure 14
<p>A visualization of model prediction results, i.e., vehicles and roads: (<b>a</b>) nighttime; and (<b>b</b>) daytime. Ego Vehicle (Red) and Predicted Vehicles (Blue).</p>
Full article ">
18 pages, 9657 KiB  
Article
Research on Digital Terrain Construction Based on IMU and LiDAR Fusion Perception
by Chen Huang, Yiqi Wang, Xiaoqiang Sun and Shiyue Yang
Sensors 2025, 25(1), 15; https://doi.org/10.3390/s25010015 - 24 Dec 2024
Viewed by 318
Abstract
To address the shortcomings of light detection and ranging (LiDAR) sensors in extracting road surface elevation information in front of a vehicle, a scheme for digital terrain construction based on the fusion of an Inertial Measurement Unit (IMU) and LiDAR perception is proposed. [...] Read more.
To address the shortcomings of light detection and ranging (LiDAR) sensors in extracting road surface elevation information in front of a vehicle, a scheme for digital terrain construction based on the fusion of an Inertial Measurement Unit (IMU) and LiDAR perception is proposed. First, two sets of sensor coordinate systems were configured, and the parameters of LiDAR and IMU were calibrated. Then, a terrain construction system based on the fusion perception of IMU and LiDAR was established, and improvements were made to the state estimation and mapping architecture. Terrain construction experiments were conducted in an academic setting. Finally, based on the output information from the terrain construction system, a moving average-like algorithm was designed to process point cloud data and extract the road surface elevation information at the vehicle’s trajectory position. By comparing the extraction effects of four different sliding window widths, the 4 cm width sliding window, which yielded the best results, was ultimately selected, making the extracted road surface elevation information more accurate and effective. Full article
(This article belongs to the Section Radar Sensors)
Show Figures

Figure 1

Figure 1
<p>LiDAR coordinate system and IMU coordinate system.</p>
Full article ">Figure 2
<p>LiDAR and IMU calibration diagram.</p>
Full article ">Figure 3
<p>Overall architecture of state estimation.</p>
Full article ">Figure 4
<p>Road segment with speed bump on campus.</p>
Full article ">Figure 5
<p>Point cloud map of road segment with speed bump.</p>
Full article ">Figure 6
<p>Point cloud map of road segment with speed bump generated using improved architecture.</p>
Full article ">Figure 7
<p>Uneven road surface used for the formal experiment.</p>
Full article ">Figure 8
<p>Point cloud map generated using the original architecture.</p>
Full article ">Figure 9
<p>Improved architecture-generated point cloud maps.</p>
Full article ">Figure 10
<p>Effect of the Passthrough algorithm.</p>
Full article ">Figure 11
<p>Flowchart of the Passthrough algorithm.</p>
Full article ">Figure 12
<p>Top view of point cloud data for the front tire trajectory position.</p>
Full article ">Figure 13
<p>Road surface elevation information generated using moving average-like algorithm.</p>
Full article ">Figure 14
<p>Road surface elevation information generated using Gaussian filter algorithm.</p>
Full article ">
20 pages, 6270 KiB  
Article
Initial Pose Estimation Method for Robust LiDAR-Inertial Calibration and Mapping
by Eun-Seok Park , Saba Arshad and Tae-Hyoung Park
Sensors 2024, 24(24), 8199; https://doi.org/10.3390/s24248199 - 22 Dec 2024
Viewed by 459
Abstract
Handheld LiDAR scanners, which typically consist of a LiDAR sensor, Inertial Measurement Unit, and processor, enable data capture while moving, offering flexibility for various applications, including indoor and outdoor 3D mapping in fields such as architecture and civil engineering. Unlike fixed LiDAR systems, [...] Read more.
Handheld LiDAR scanners, which typically consist of a LiDAR sensor, Inertial Measurement Unit, and processor, enable data capture while moving, offering flexibility for various applications, including indoor and outdoor 3D mapping in fields such as architecture and civil engineering. Unlike fixed LiDAR systems, handheld devices allow data collection from different angles, but this mobility introduces challenges in data quality, particularly when initial calibration between sensors is not precise. Accurate LiDAR-IMU calibration, essential for mapping accuracy in Simultaneous Localization and Mapping applications, involves precise alignment of the sensors’ extrinsic parameters. This research presents a robust initial pose calibration method for LiDAR-IMU systems in handheld devices, specifically designed for indoor environments. The research contributions are twofold. Firstly, we present a robust plane detection method for LiDAR data. This plane detection method removes the noise caused by mobility of scanning device and provides accurate planes for precise LiDAR initial pose estimation. Secondly, we present a robust planes-aided LiDAR calibration method that estimates the initial pose. By employing this LiDAR calibration method, an efficient LiDAR-IMU calibration is achieved for accurate mapping. Experimental results demonstrate that the proposed method achieves lower calibration errors and improved computational efficiency compared to existing methods. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

Figure 1
<p>LiDAR based mapping using (<b>a</b>) LiDAR-IMU calibration method: Error-free mapping, and (<b>b</b>) Without LiDAR-IMU calibration method: Mapping error due to drift, highlighted in yellow circle. The colors in each map represents the intensity of LiDAR point cloud.</p>
Full article ">Figure 2
<p>Overall framework of the proposed initial pose estimation method for robust LiDAR-IMU calibration. Different colors in voxelization shows the intensity of LiDAR points in each voxel. The extracted planes are represented with yellow and green color while red color points indicate noise.</p>
Full article ">Figure 3
<p>Robust plane detection method.</p>
Full article ">Figure 4
<p>Robust plane extraction through refinement. (<b>a</b>) Voxels containing edges and noise have low plane scores due to large distances and high variance represented as red color normal vector while those with high plane scores are represented with blue. (<b>b</b>) The refinement process enables the effective separation and removal of areas containing edges and noise.</p>
Full article ">Figure 5
<p>LiDAR calibration method.</p>
Full article ">Figure 6
<p>IMU downsampling.</p>
Full article ">Figure 7
<p>Qualitative Comparison of the proposed method with the benchmark plane detection algorithms.</p>
Full article ">Figure 8
<p>Top view of LiDAR data. (<b>a</b>) LiDAR raw data before calibration. (<b>b</b>) LiDAR data after calibration using the proposed method.</p>
Full article ">Figure 9
<p>Performance comparison in terms of (<b>a</b>) roll and (<b>b</b>) pitch errors in the VECtor dataset.</p>
Full article ">Figure 10
<p>Performance comparison in terms of the (<b>a</b>) mapping result using LI-init and (<b>b</b>) mapping result using LI-init+Proposed.</p>
Full article ">
27 pages, 90792 KiB  
Article
Integration of Structural Characteristics from GEDI Waveforms for Improved Forest Type Classification
by Mary M. McClure, Satoshi Tsuyuki and Takuya Hiroshima
Remote Sens. 2024, 16(24), 4776; https://doi.org/10.3390/rs16244776 - 21 Dec 2024
Viewed by 385
Abstract
Forest types correspond to differences in structural characteristics and species composition that influence biomass and biodiversity values, which are essential measurements for ecological monitoring and management. However, differentiating forest types in tropical regions remains a challenge. This study aimed to improve forest type [...] Read more.
Forest types correspond to differences in structural characteristics and species composition that influence biomass and biodiversity values, which are essential measurements for ecological monitoring and management. However, differentiating forest types in tropical regions remains a challenge. This study aimed to improve forest type extent mapping by combining structural information from discrete full-waveform LiDAR returns with multitemporal images. This study was conducted in a tropical forest region over complex terrain in north-eastern Tanzania. First, structural classes were generated by applying time-series clustering algorithms. The results showed four different structural clusters corresponding to forest types, montane–humid forest, montane–dry forest, submontane forest, and non-forest, when using the Kshape algorithm. Kshape considers the shape of the full-sequence LiDAR waveform, requiring little preprocessing. Despite the overlap amongst the original clusters, the averages of structural characteristics were significantly different across all but five metrics. The labeled clusters were then further refined and used as training data to generate a wall-to-wall forest cover type map by classifying biannual images. The highest-performing model was a KNN model with 13 spectral and 3 terrain features achieving 81.7% accuracy. The patterns in the distributions of forest types provide better information from which to adapt forest management, particularly in forest–non-forest transitional zones. Full article
(This article belongs to the Section Environmental Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Location and extent of the study area within Same District, Tanzania. The red border represents the forest reserve boundary provided by WDPA [<a href="#B18-remotesensing-16-04776" class="html-bibr">18</a>]. The yellow points represent the available geolocated footprints within the bounding box of the reserve after filtering by quality flags, while the red points represent the locations of forest inventory plots.</p>
Full article ">Figure 2
<p>Dunn’s test results showing the significant and nonsignificant (denoted by an X) features between cluster pairs. The full table of the Kruskal–Wallis test results can be found in <a href="#app3-remotesensing-16-04776" class="html-app">Appendix C</a>. Feature explanations can be found in <a href="#app1-remotesensing-16-04776" class="html-app">Appendix A</a> (<a href="#remotesensing-16-04776-t0A1" class="html-table">Table A1</a>).</p>
Full article ">Figure 3
<p>Correlation heat map among structural features. The plot shows high correlations between waveform metrics such as wavelength, wave distance, and centroid with derived vegetation metrics, such as the relative heights. Plant cover, volume, metrics, and rh100 showed a strong negative correlation with ‘peak’ amplitude, but a positive correlation with ‘TotalAmp’, likely caused by the scattering of signal in denser, taller vegetation.</p>
Full article ">Figure 4
<p>Classified forest map of Kwizu Forest reserve. The forest type classification map was resampled to 100 m resolution.</p>
Full article ">Figure 5
<p>Individual waveforms (blue) and mean waveform (red) of each cluster. The mean waveform was the average amplitude value at each time point for all the observations in the cluster. Cluster 0 was associated with montane–dry forest, cluster 1 with non-forest, 2 with montane–humid forest, and 3 with submontane forest.</p>
Full article ">Figure A1
<p>The locations of filtered GEDI footprints overlayed on the study area in light blue, the green points represent the footprints that were extracted for the training data. The locations of the field inventory are shown in red. The underlying image is an RGB composite image from January 2023 at 4.77 m resolution from Norway’s International Climate and Forest Initiative (NICFI), available in Google Earth Engine (GEE). To the right is a map of elevation gradient overlaid on hill shade to better illustrate the topographic complexity.</p>
Full article ">
Back to TopTop