[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
μ-FTIR, μ-Raman, and SERS Analysis of Amide I Spectral Region in Oral Biofluid Samples during Orthodontic Treatment
Next Article in Special Issue
Data Center Traffic Prediction Algorithms and Resource Scheduling
Previous Article in Journal
Beyond PRNU: Learning Robust Device-Specific Fingerprint for Source Camera Identification
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Drone-Based Environmental Monitoring and Image Processing Approaches for Resource Estimates of Private Native Forest

by
Sanjeev Kumar Srivastava
1,
Kah Phooi Seng
2,3,*,
Li Minn Ang
1,
Anibal ‘Nahuel’ A. Pachas
4 and
Tom Lewis
1
1
School of Science Technology and Engineering, University of the Sunshine Coast, Sippy Downs, QLD 4556, Australia
2
School of AI and Advanced Computing, Xi’an Jiaotong Liverpool University, Suzhou 215000, China
3
School of Computer Science, Queensland University of Technology, Brisbane City, QLD 4000, Australia
4
Department of Agriculture and Fisheries, Queensland Government, 1 Cartwright Road, Gympie, QLD 4570, Australia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(20), 7872; https://doi.org/10.3390/s22207872
Submission received: 29 September 2022 / Revised: 7 October 2022 / Accepted: 11 October 2022 / Published: 17 October 2022
Figure 1
<p>Location of private native forest sites selected for the study and the drone flight path used to collect images. Sites were located near Ravensbourne (site 1) and Esk (site 2). Background true colour images are from different remote-sensing satellites available in ESRI base maps.</p> ">
Figure 2
<p>Methodological framework used in this study.</p> ">
Figure 3
<p>Location of the first site selected for analysing drone data and the orthomosaic of the site.</p> ">
Figure 4
<p>Separation of the orthomosaic into red, green and blue spectrum.</p> ">
Figure 5
<p>Normalisation of red, green and blue spectrum and calculation of various indices (see <a href="#sensors-22-07872-t002" class="html-table">Table 2</a>) from normalised bands.</p> ">
Figure 6
<p>Identification of an index and its reclassification to extract trees and their differentiation into three potential classes, that may represent different tree species.</p> ">
Figure 7
<p>Generation of canopy height model allowing extraction of height information for different trees.</p> ">
Figure 8
<p>Location of site 2 selected for analysing drone data.</p> ">
Figure 9
<p>Identification of an index and its reclassification to extract trees and their differentiation into three classes.</p> ">
Figure 10
<p>Generating a canopy height model for extraction of height information for different trees.</p> ">
Figure 11
<p>Location of field-measured tree height data points.</p> ">
Figure 12
<p>Relationship and statistical test between field-measured and drone-derived tree heights.</p> ">
Versions Notes

Abstract

:
This paper investigated the utility of drone-based environmental monitoring to assist with forest inventory in Queensland private native forests (PNF). The research aimed to build capabilities to carry out forest inventory more efficiently without the need to rely on laborious field assessments. The use of drone-derived images and the subsequent application of digital photogrammetry to obtain information about PNFs are underinvestigated in southeast Queensland vegetation types. In this study, we used image processing to separate individual trees and digital photogrammetry to derive a canopy height model (CHM). The study was supported with tree height data collected in the field for one site. The paper addressed the research question “How well do drone-derived point clouds estimate the height of trees in PNF ecosystems?” The study indicated that a drone with a basic RGB camera can estimate tree height with good confidence. The results can potentially be applied across multiple land tenures and similar forest types. This informs the development of drone-based and remote-sensing image-processing methods, which will lead to improved forest inventories, thereby providing forest managers with recent, accurate, and efficient information on forest resources.

1. Introduction

There are approximately 2 million hectares of potentially harvestable private native forest (PNF) in southern Queensland. However, there is little information on the productive state (in terms of timber production) of this resource. Field surveys suggest that large parts of this resource are in a relatively unproductive condition. However, some works have shown that management (e.g., tree thinning) can greatly enhance the productive value of the native forests and that remote-sensing options can provide useful information on the heights and species within native forest environments. There is a need for forest inventory work to determine where forest management might be best targeted. If well managed, this resource has great potential for helping maintain or grow the hardwood timber industry in Queensland. Image processing and remote sensing have the potential to increase cost-effectiveness and time efficiency and reduce uncertainty compared with the traditional method of generating forest inventory data from sampling plots [1,2,3]. Accurate, spatially detailed forestry inventories can be generated from cost-effective, near real-time data sets [4,5,6].
In the literature, four remote sensing approaches have generally been used to monitor forest environments: (1) airborne and terrestrial light detection and ranging (LiDAR); (2) interferometric radar; (3) high-resolution multispectral images, and (4) photogrammetry [7,8,9]. LiDAR (light detection and ranging) data and high-resolution remote-sensing images (e.g., photogrammetry) can be used to develop a canopy point cloud. Where the density of this point cloud is high enough (i.e., high spatial resolution), this allows for the determination of the number of trees per hectare and their canopy heights. However, flying over potential forests of interest to obtain LiDAR data is cost-prohibitive, and while high-resolution satellite images are useful for giving an indication of the overall canopy height, this method of photogrammetry does not allow the determination of the number of trees per hectare (point clouds derived are not of sufficient density) in the forest. Motion photogrammetry using unmanned aerial vehicles (i.e., drones) can produce high-density point clouds to allow the estimation of productive tree crowns. However, to the best of our knowledge, this method has not been validated in subtropical eucalypt forests.
Forest inventory data collection is enhanced by recent developments in UAV image collection, and products using digital photogrammetry are cost-effective options for forest structure profiling [10]). There are several examples of UAV-supported remote-sensing techniques that have been used to study forests [11,12]. Several studies have demonstrated the accuracy of aerial laser survey (ALS) or LiDAR for providing forest inventory data [13,14,15,16,17] and its efficiency compared with field measurements [18]. Yet ALS can be cost-prohibitive, especially when analysing temporal change. The high cost associated with ALS can limit the spatial extent and temporal frequency of data collection [19,20,21]. Drone-based approaches have several advantages for forest monitoring. Lower flight altitudes are able to give high spatial–temporal resolution, data acquisition frequency can be high due to the relatively low operating costs, and the algorithms associated with UAV-derived products are robust and provide accurate elevation data.
This paper investigates the utility of drone-based monitoring and image processing approaches to assist with forest inventory in Queensland native forests. The research aims to build capabilities to carry out forest inventory far more efficiently without the need to rely on laborious field assessments. In this paper, we used a number of existing private native forest sites where field inventory data was available. We investigated the utility of drone-based remote-sensing and image-processing data for the collection of forest inventory information. The findings of this study will be highly beneficial for various agencies working on managing natural and commercial forests. This paper is structured as follows. Section 2 discusses some previous works on drone-based approaches for environmental and forest monitoring. Section 3 describes the methodology and data collection approaches used for the investigation. Section 4 discusses the results and the interpretation of the results. Some concluding remarks are given in Section 5.

2. Recent Works on Drone-Based Approaches for Environmental Monitoring

The objective of this section is to provide background information before we discuss our proposed approach for drone-based environmental monitoring and image processing to assist with forest inventory applied to the private native forests in southern Queensland. Table 1 shows a summary of the works. We give some examples for two categories that commonly use drone-based environmental monitoring: (1) land monitoring and identification and (2) aquatic/marine monitoring and identification. The work proposed in this paper is for land-based monitoring and identification, with a focus on utilizing drone-based imagery for forest inventory.

3. Methodology and Data Collection

3.1. Site Selection for Drone Monitoring

This research used two privately owned sites where the native forest is managed for timber and grazing production. The drone data were collected for different sites in the period from August to November 2020 and the field–forest inventory data were collected at the same time. The research utilised point clouds derived from drone flights over the selected sites. The following software packages were used for extracting information: (1) ArcGIS Pro/ArcGIS; (2) R-statistics; (3) ENVI; and (4) Metashape.
Images were collected using a Phantom 4 pro drone with 4Kcamera in red, green and blue spectrum during August-September 2020. The drone was flown in a Double-Grid PIX4D Capture pattern, at an altitude of 60 m above ground level (AGL), with a 65° camera angle and 80% lateral and longitudinal image overlap. Flights were carried out in accordance with Australia’s Civil Aviation Safety Authority (CASA) regulations for safe operation of unmanned aerial vehicles (UAVs) and by suitably qualified personnel in clear weather conditions with maximum winds not exceeding 20 kph. Images were stored on a 32 GB micro SD card mounted on the drone and subsequently used for analysis.
Figure 1 shows the location of the private native forest sites selected for the study and the flight path of the drone for collecting the image data. The novel field assessment mapped the individual tree locations using a Trimble GeoXH Geoexplorer, a differential global navigation satellite system (GNSS) with up to 10 cm accuracy. Individual trees were located within existing permanent monitoring plots (generally 40 × 40 m) distributed within different sites selected for the study. For each tree ≥10 cm, diameter at breast height (DBH) and tree height were recorded.

3.2. Image Analysis of Drone Images

The RGB images collected from the drone were first analysed in Agisoft Metashape software. With multiple overlapping 2d drone images, using structure from motion (SfM), a digital photogrammetry technique, a point cloud and digital surface model (DSM) were generated together with an orthomosaic (Figure 2).
The SfM algorithm is a four-step process where a point of interest is located in each overlapping image. This is automated in software packages such as Metashape. The next step involved finding candidate correspondences with matching descriptors for each interest point. In the third step, a geometric verification of correspondences is performed. This uses the RANSAC algorithm to exclude outliers while matching features. The final step finds a solution for 3d points to minimise reprojection error. This is also referred as bundle adjustment.
An orthomosaic from drone images is a geometrically corrected mosaic that is colour balanced and rectified for distortions due to elevation to create a seamless image of an area. Similarly, the point cloud is a cloud of points with horizontal (x and y) as well as vertical (height) coordinates. Point clouds can easily be interpolated to create a digital raster surface of elevation values.
The RGB orthomosaic was separated into red (R), green (G), and blue (B) spectral bands that were further normalised using the following equations:
Normalised   red   ( r ) = R R + G + B
Normalised   green   ( g ) = G R + G + B
Normalised   blue   ( b ) = B R + G + B
Using the normalised red, green, and blue bands, a number of colour indices were calculated. Table 2 shows a summary of the colour indices, formulas, and references thatwhich were used in the models.
We used ArcGIS Pro 2.8 to calculate all the indices. For this study, we used normalised green-red difference index (NGBDI) to separate vegetation from other features. On visual inspection of all the indices, the normalised green-red difference index (NGBDI) was found best for separating individual trees. Considering this, we used NGBDI to separate trees from other features. Once vegetation was separated, a separate cleaned polygon was generated for vegetation.
We further used reclassified NGBDI images to generate random locations on the ground surface to generate a digital elevation model (DEM). Using the following equation, we calculated the canopy height model (CHM):
CHM = DSM − DEM
Tree height and tree diameter at breast height (DBH) were measured at 116 locations for 1 of the study areas. The data were collected for site 2 (Figure 1) during field trips in September and November 2020.
To test the accuracy of the tree heights estimated from the drone images, we created a buffer of 1 m around each field data collection location and picked the maximum value from the canopy height model (CHM) within this zone. To test the statistical relationship between drone-derived and the height measured in field, we used a Pearson’s correlation test using the ggpubr library available with R statistics software.

4. Results and Discussion

The orthomosaic created for the different sites clearly showed individual trees, especially near the flight path of the drone. Figure 3 shows the location of the first site selected for analysing drone data and the orthomosaic of the site. When the orthomosaic was separated into RGB bands, variation in information content was noticed that was enhanced with the calculation of indices. Figure 4 shows the separation of the orthomosaic into red, green, and blue spectra. The green band showed all the vegetated areas more clearly, while the vegetation was not clearly separated in the blue and green bands (Figure 4).
The visual analysis of indices clearly indicated variation in information content that could be further utilised for extracting multiple features such as vegetation types, bare soil, or grass area. For this study, we selected NGBVI to separate vegetation. Figure 5 shows the normalisation of the red, green, and blue spectra and the calculations of various indices from normalised bands.
Figure 6 shows the identification of an index and its reclassification to extract trees and their differentiation into three classes. For the first site, the reclassification of NGBVI clearly differentiated trees from other features. Variation between three vegetation types based on NGBVI was noted, but this study focuses only on separating trees and estimating their height. The separation of trees enabled the extraction of tree crown areas, which were then converted to simplified polygons with a separate identification. Figure 7 shows the generation of the canopy height model to allow for the extraction of height information for different trees. The polygons created for each tree were further used to perform a zonal operation on CHM to acquire detailed statistics of the height of each tree. Some further results for a second site are shown in Figure 8, Figure 9 and Figure 10. For this site, tree height statistics and crown area were derived, and tree heights were compared with those assessed in the field (Figure 11). There was a strong relationship between field-measured heights and those derived from the CHM (p < 0.0001, Figure 12).

5. Conclusions

This paper has investigated drone-based and image processing approaches for forest inventory monitoring. The work has shown that vegetated areas and potentially the species contained within can be calculated from indices derived from red, green, and blue bands of orthomosaic images. Furthermore, using CHM data, it was possible to calculate height statistics for vegetated areas. The combination of RGB-derived indices and a canopy height model (CHM) enabled obtaining detailed inventory information for individual trees within PNFs. It was possible to calculate height statistics for individual trees with reasonable accuracy, although this would be more challenging in dense forests with overlapping tree crowns. A few plots were affected by tree shadow, which interfered with other information, although this was minimised to a greater extent with the calculation of indices. This exploratory study suggested that using a combination of different indices and the canopy height model, it will be possible to apply machine and deep learning algorithms to separate vegetation types or species. Future work will investigate this approach and also estimate forest biomass using height information and the area occupied by different trees.

Author Contributions

Conceptualization, S.K.S. and A.‘N.A.P.; methodology, S.K.S.; resources, K.P.S.; data curation, S.K.S., L.M.A. and A.‘N.A.P.; writing—original draft preparation, S.K.S., K.P.S., T.L. and L.M.A.; writing—review and editing, S.K.S., K.P.S., T.L. and L.M.A. All authors have read and agreed to the published version of the manuscript.

Funding

This study is funded by Queensland Department of Agriculture and Fisheries.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

Authors declare no conflict of interest.

References

  1. Lewis, T.; Menzies, T.; Brawner, J.T.; Venn, T.; Francis, B.; Ryan, S.; Schulke, B.; Cameron, N.; Kelly, A. Improving Productivity of the Private Native Forest Resource in Southern Queensland and Northern New South Wales. Available online: https://www.fwpa.com.au/resources/reports/resources.html (accessed on 11 April 2022).
  2. Dash, J.P.; Watt, M.S.; Bhandari, S.; Watt, P. Characterising forest structure using combinations of airborne laser scanning data, RapidEye satellite imagery and environmental variables. Forestry 2015, 89, 159–169. [Google Scholar] [CrossRef] [Green Version]
  3. White, J.C.; Coops, N.C.; Wulder, M.A.; Vastaranta, M.; Hilker, T.; Tompalski, P. Remote sensing technologies for enhancing forest inventories: A review. Can. J. Remote Sens. 2016, 42, 619–641. [Google Scholar] [CrossRef] [Green Version]
  4. Wulder, M.A.; Coops, N.C.; Hudak, A.T.; Morsdorf, F.; Nelson, R.; Newnham, G.; Vastaranta, M. Status and prospects for LiDAR remote sensing of forested ecosystems. Can. J. Remote Sens. 2013, 39, S1–S5. [Google Scholar] [CrossRef] [Green Version]
  5. Groot, A.; Cortini, F.; Wulder, M.A. Crown-fibre attribute relationships for enhanced forest inventory: Progress and prospects. For. Chron. 2015, 91, 266–279. [Google Scholar] [CrossRef] [Green Version]
  6. Stone, C.; Osborn, J. Deployment and Integration of Cost-Effective High Resolution Remotely Sensed Data for the Australian Forest Industry. Melb. For. Wood Prod. Aust. 2017. Available online: https://www.fwpa.com.au/resources/resources/1461-deployment-and-integration-of-cost-effective-high-spatial-resolution-remotely-sensed-data-for-the-australian-forestry-industry.html (accessed on 28 September 2022).
  7. Yu, X.; Hyyppä, J.; Karjalainen, M.; Nurminen, K.; Karila, K.; Vastaranta, M.; Kankare, V.; Kaartinen, H.; Holopainen, M.; Honkavaara, E.; et al. Comparison of laser and stereo optical, SAR and InSAR point clouds from air-and space-borne sources in the retrieval of forest inventory attributes. Remote Sens. 2015, 7, 15933–15954. [Google Scholar] [CrossRef] [Green Version]
  8. Stone, C.; Webster, M.; Osborn, J.; Iqbal, I. Alternatives to LiDAR-derived canopy height models for softwood plantations: A review and example using photogrammetry. Aust. For. 2016, 79, 271–282. [Google Scholar] [CrossRef]
  9. Olive, K.; Lewis, T.; Ghaffariyan, M.R.; Srivastava, S.K. Comparing canopy height estimates from satellite-based photogrammetry, airborne laser scanning and field measurements across Australian production and conservation eucalypt forests. J. For. Res. 2020, 25, 108–112. [Google Scholar] [CrossRef]
  10. Iglhaut, J.; Cabo, C.; Puliti, S.; Piermattei, L.; O’Connor, J.; Rosette, J. Structure from Motion Photogrammetry in Forestry: A Review. Curr. For. Rep. 2019, 5, 155–168. [Google Scholar] [CrossRef] [Green Version]
  11. Goodbody, T.R.H.; Coops, N.C.; White, J.C. Digital aerial photogrammetry for updating area-based forest inventories: A review of opportunities, challenges, and future directions. Curr. For. Rep. 2019, 5, 55–75. [Google Scholar] [CrossRef] [Green Version]
  12. Alejandro, P.; Ruiz Fernández, L.Á.; Gómez, C.; Pascual, C.; Álvarez Taboada, M.F.; Tanase, M.; Valbuena, R.; Montes Pita, F.; Hermosilla, T. Remote Sensing for the Spanish Forests in the 21st Century: A Review of Advances, Needs, and Opportunities. For. Syst. 2019, 28, 1–33. Available online: https://usc.primo.exlibrisgroup.com/permalink/61USC_INST/3ct1hk/cdi_dialnet_primary_oai_dialnet_unirioja_es_ART0001340628 (accessed on 1 October 2022).
  13. Lim, K.S.; Treitz, P.M. Estimation of above ground forest biomass from airborne discrete return laser scanner data using canopy-based quantile estimators. Scand. J. For. Res. 2004, 19, 558–570. [Google Scholar] [CrossRef]
  14. Kumar, S.S.; Tom, L.; Linda, B.; Stuart, P. Spatial databases and techniques to assist with prescribed fire management in the south-east Queensland bioregion. Int. J. Wildland Fire 2021, 30, 90–111. [Google Scholar]
  15. Roberts, S.D.; Dean, T.J.; Evans, D.L.; McCombs, J.W.; Harrington, R.L.; Glass, P.A. Estimating individual tree leaf area in loblolly pine plantations using LiDAR-derived measurements of height and crown dimensions. For. Ecol. Manag. 2005, 213, 54–70. [Google Scholar] [CrossRef]
  16. Popescu, S.C.; Zhao, K. A voxel-based lidar method for estimating crown base height for deciduous and pine trees. Remote Sens. Environ. 2008, 112, 767–781. [Google Scholar] [CrossRef]
  17. Næsset, E. Practical large-scale forest stand inventory using a small-footprint airborne scanning laser. Scand. J. For. Res. 2004, 19, 164–179. [Google Scholar] [CrossRef]
  18. Andersen, H.-E.; Reutebuch, S.E.; McGaughey, R.J. A rigorous assessment of tree height measurements obtained using airborne lidar and conventional field methods. Can. J. Remote Sens. 2006, 32, 355–366. [Google Scholar] [CrossRef]
  19. Lu, D.; Chen, Q.; Wang, G.; Liu, L.; Li, G.; Moran, E. A survey of remote sensing-based aboveground biomass estimation methods in forest ecosystems. Int. J. Digit. Earth 2016, 9, 63–105. [Google Scholar] [CrossRef]
  20. Pearse, G.D.; Dash, J.P.; Persson, H.J.; Watt, M.S. Comparison of high-density LiDAR and satellite photogrammetry for forest inventory. ISPRS J. Photogramm. Remote Sens. 2018, 142, 257–267. [Google Scholar] [CrossRef]
  21. Krause, S.; Sanders, T.G.M.; Mund, J.-P.; Greve, K. UAV-Based Photogrammetric Tree Height Measurement for Intensive Forest Monitoring. Remote Sens. 2019, 11, 758. [Google Scholar] [CrossRef] [Green Version]
  22. Padró, J.C.; Cardozo, J.; Montero, P.; Ruiz-Carulla, R.; Alcañiz, J.M.; Serra, D.; Carabassa, V. Drone-based identification of erosive processes in open-pit mining restored areas. Land 2022, 11, 212. [Google Scholar] [CrossRef]
  23. Reckling, W.; Mitasova, H.; Wegmann, K.; Kauffman, G.; Reid, R. Efficient Drone-Based Rare Plant Monitoring Using a Species Distribution Model and AI-Based Object Detection. Drones 2021, 5, 110. [Google Scholar] [CrossRef]
  24. Thomazella, R.; Castanho, J.E.; Dotto, F.R.; Júnior, O.R.; Rosa, G.H.; Marana, A.N.; Papa, J.P. Environmental monitoring using drone images and convolutional neural networks. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 8941–8944. [Google Scholar]
  25. Tóth, V.Z.; Grósz, J.; Ladányi, M.; Jung, A. A new lake algae detection method supported by a drone-based multispectral camera. Lakes Reserv. Res. Manag. 2021, 26, e12377. [Google Scholar] [CrossRef]
  26. Duan, Z.; Li, Y.; Wang, X.; Wang, J.; Brydegaard, M.; Zhao, G.; Svanberg, S. Drone-based fluorescence lidar systems for vegetation and marine environment monitoring. In EPJ Web of Conferences; EDP Sciences: Ulys, France, 2020; Volume 237, p. 07013. [Google Scholar]
  27. Ventura, D.; Bruno, M.; Lasinio, G.J.; Belluscio, A.; Ardizzone, G. A low-cost drone-based application for identifying and mapping of coastal fish nursery grounds. Estuar. Coast. Shelf Sci. 2016, 171, 85–98. [Google Scholar] [CrossRef]
Figure 1. Location of private native forest sites selected for the study and the drone flight path used to collect images. Sites were located near Ravensbourne (site 1) and Esk (site 2). Background true colour images are from different remote-sensing satellites available in ESRI base maps.
Figure 1. Location of private native forest sites selected for the study and the drone flight path used to collect images. Sites were located near Ravensbourne (site 1) and Esk (site 2). Background true colour images are from different remote-sensing satellites available in ESRI base maps.
Sensors 22 07872 g001
Figure 2. Methodological framework used in this study.
Figure 2. Methodological framework used in this study.
Sensors 22 07872 g002
Figure 3. Location of the first site selected for analysing drone data and the orthomosaic of the site.
Figure 3. Location of the first site selected for analysing drone data and the orthomosaic of the site.
Sensors 22 07872 g003
Figure 4. Separation of the orthomosaic into red, green and blue spectrum.
Figure 4. Separation of the orthomosaic into red, green and blue spectrum.
Sensors 22 07872 g004
Figure 5. Normalisation of red, green and blue spectrum and calculation of various indices (see Table 2) from normalised bands.
Figure 5. Normalisation of red, green and blue spectrum and calculation of various indices (see Table 2) from normalised bands.
Sensors 22 07872 g005
Figure 6. Identification of an index and its reclassification to extract trees and their differentiation into three potential classes, that may represent different tree species.
Figure 6. Identification of an index and its reclassification to extract trees and their differentiation into three potential classes, that may represent different tree species.
Sensors 22 07872 g006
Figure 7. Generation of canopy height model allowing extraction of height information for different trees.
Figure 7. Generation of canopy height model allowing extraction of height information for different trees.
Sensors 22 07872 g007
Figure 8. Location of site 2 selected for analysing drone data.
Figure 8. Location of site 2 selected for analysing drone data.
Sensors 22 07872 g008
Figure 9. Identification of an index and its reclassification to extract trees and their differentiation into three classes.
Figure 9. Identification of an index and its reclassification to extract trees and their differentiation into three classes.
Sensors 22 07872 g009
Figure 10. Generating a canopy height model for extraction of height information for different trees.
Figure 10. Generating a canopy height model for extraction of height information for different trees.
Sensors 22 07872 g010
Figure 11. Location of field-measured tree height data points.
Figure 11. Location of field-measured tree height data points.
Sensors 22 07872 g011
Figure 12. Relationship and statistical test between field-measured and drone-derived tree heights.
Figure 12. Relationship and statistical test between field-measured and drone-derived tree heights.
Sensors 22 07872 g012
Table 1. Some recent works on drone-based approaches for environmental monitoring.
Table 1. Some recent works on drone-based approaches for environmental monitoring.
Domain AreaYearEnvironmental ApplicationTechniques AppliedReference
Land monitoring and identification2022Identification of water erosion in mining restored areasThe generation of a digital elevation model (DEM) from drone images and photogrammetric processingPadro et al., 2022 [22]
2021Monitoring of rare plant speciesNeural network object detector (YOLO) and darknet neural network frameworkReckling et al., 2021 [23]
2018Monitoring of natural/protected reserves from illegal activitiesCNN (VGG16, VGG19) and transfer learning to classify four terrain classes: water, deforesting, forest and buildingsThomazella et al., 2018 [24]
Aquatic/marine monitoring and identification2021Detection and quantification of algal in water ecosystemsImage analysis of drone-based multispectral imageryToth et al., 2021 [25]
2020Drone-based fluorosensor for marine environment monitoringAnalysis of hyperspectral lidar data and fluorescence spectral recordings for vegetation profilingDuan et al., 2020 [26]
2016Mapping of coastal fish nursery grounds and marine habitatsImage segmentation tools (ArcGIS, MultiSpec, eCognition Developer) to classify marine terrain classes: coarse sand, fine sand, leaves, matte, shallow rock, deep rock Ventura et al., 2016 [27]
Table 2. Colour indices, formulas, and references.
Table 2. Colour indices, formulas, and references.
IndicesFormula
1Visible Atmospherically Resistant Index (VARI) (Gitelson et al. 2002)(g − r)/(g + r − b)
2Excess Green Vegetation Index (ExG) (Woebbecke et al. 1995)2g − r − b
3Excess Red Vegetation Index (ExR) (Meyer et al. 2008)1.4r − g
4Excess Blue Vegetation Index (ExB) (Mao et al. 2003)1.4b − g
5Excess Green minus Excess Red Vegetation Index (ExGR) (Neto 2004)ExG − ExR
6Normalized Green-Red Difference Index (NGRDI) (Tucker 1979)(g − r)/(g + r)
7Normalized Green-Red Difference Index (NGBDI) (Tucker 1979)(g − b)/(g + b)
8Modified Green Red Vegetation Index (MGRVI) (Tucker 1979)(g2 − r2)/(g2 + r2)
9Woebbecke Index (WI) (Woebbecke et al. 1995)(g − b)/(r − g)
10Kawashima Index (IKAW) (Kawashima and Nakatani 1998)(r − b)/(r + b)
11Green Leaf Algorithm (GLA) (Louhaichi et al. 2001)(2g − r − b)/(2g + r + b)
12Red Green Blue Vegetation Index (RGBVI) (Bendig et al. 2015)(g2 − b × r)/(g2 + b × r)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Srivastava, S.K.; Seng, K.P.; Ang, L.M.; Pachas, A.‘N.A.; Lewis, T. Drone-Based Environmental Monitoring and Image Processing Approaches for Resource Estimates of Private Native Forest. Sensors 2022, 22, 7872. https://doi.org/10.3390/s22207872

AMA Style

Srivastava SK, Seng KP, Ang LM, Pachas A‘NA, Lewis T. Drone-Based Environmental Monitoring and Image Processing Approaches for Resource Estimates of Private Native Forest. Sensors. 2022; 22(20):7872. https://doi.org/10.3390/s22207872

Chicago/Turabian Style

Srivastava, Sanjeev Kumar, Kah Phooi Seng, Li Minn Ang, Anibal ‘Nahuel’ A. Pachas, and Tom Lewis. 2022. "Drone-Based Environmental Monitoring and Image Processing Approaches for Resource Estimates of Private Native Forest" Sensors 22, no. 20: 7872. https://doi.org/10.3390/s22207872

APA Style

Srivastava, S. K., Seng, K. P., Ang, L. M., Pachas, A. ‘N. A., & Lewis, T. (2022). Drone-Based Environmental Monitoring and Image Processing Approaches for Resource Estimates of Private Native Forest. Sensors, 22(20), 7872. https://doi.org/10.3390/s22207872

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop