Next Article in Journal
Spatial-Temporal Evolution and Influencing Factors Analysis of Ecosystem Services Value: A Case Study in Sunan Canal Basin of Jiangsu Province, Eastern China
Previous Article in Journal
ERA5 Reanalysis for the Data Interpretation on Polarization Laser Sensing of High-Level Clouds
Previous Article in Special Issue
Individual Tree Detection in Coal Mine Afforestation Area Based on Improved Faster RCNN in UAV RGB Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tree Species Classifications of Urban Forests Using UAV-LiDAR Intensity Frequency Data

1
State Key Laboratory of Subtropical Silviculture, Zhejiang A & F University, Hangzhou 311300, China
2
Key Laboratory of Carbon Cycling in Forest Ecosystems and Carbon Sequestration of Zhejiang Province, Zhejiang A & F University, Hangzhou 311300, China
3
School of Environmental and Resources Science, Zhejiang A & F University, Hangzhou 311300, China
4
The College of Forestry, Beijing Forestry University, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Submission received: 2 November 2022 / Revised: 19 December 2022 / Accepted: 22 December 2022 / Published: 25 December 2022
(This article belongs to the Special Issue Applications of Individual Tree Detection (ITD))

Abstract

:
The accurate classification of tree species is essential for the sustainable management of forest resources and the effective monitoring of biodiversity. However, a literature review shows that most of the previous unmanned aerial vehicle (UAV) light detection and ranging (LiDAR)-based studies on fine tree species classification have used only limited intensity features, accurately identifying relatively few tree species. To address this gap, this study proposes developing a new intensity feature—intensity frequency—for the LiDAR-based fine classification of eight tree species. Intensity frequency is defined as the number of times a certain intensity value appears in the individual tree crown (ITC) point cloud. In this study, we use UAV laser scanning to obtain LiDAR data from urban forests. Intensity frequency features are constructed based on the extracted intensity information, and a random forest (RF) model is used to classify eight subtropical forest tree species in southeast China. Based on four-point cloud density sampling schemes of 100%, 80%, 50% and 30%, densities of 230 points/m2, 184 points/m2, 115 points/m2 and 69 points/m2 are obtained. These are used to analyze the effect of intensity frequency on tree species classification accuracy under four different point cloud densities. The results are shown as follows. (1) Intensity frequencies of trees are not significantly different for intraspecies (p > 0.05) values and are significantly different for interspecies (p < 0.01) values. (2) The intensity frequency features of LiDAR can be used to classify different tree species with an overall accuracy (OA) of 86.7%. Acer Buergerianum achieves a user accuracy (UA) of over 95% and a producer accuracy (PA) of over 90% for four density conditions. (3) The OA varies slightly under different point cloud densities, but the sum of correct classification trees (SCI) and PA decreases rapidly as the point cloud density decreases, while UA is less affected by density with some stability. (4) The priori feature selected by mean rank (MR) covers the top 10 posterior features selected by RF. These results show that the new intensity frequency feature proposed in this study can be used as a comprehensive and effective intensity feature for the fine classification of tree species.

Graphical Abstract

1. Introduction

The classification of tree species for research on forest inventory [1], carbon storage assessment [2], habitat [3], and ecosystem changes [4] is crucial. Remote sensing data have characteristics such as simultaneous monitoring and repeatability over large areas, which can provide landscape-scale perspectives and forest structure information [5], offering an excellent advantage for large-scale tree monitoring, mapping, and assessment. Therefore, remote sensing technology has become essential for interpreting biodiversity information. How to classify tree species from remote sensing data has been a trending research topic both in China and abroad.
Obtaining forest tree species information through remote sensing usually requires direct and indirect methods [6]. The direct methods directly classify species, community types, and abundance with data based on high spatial or spectral resolution. The spatial resolution of the data often reaches the individual tree scale or even the leaf scale to perform ITC detection [7,8,9]. Studies have shown that tree species classification accuracy of more than 80% can be achieved by using data with spatial resolution at the leaf scale and spectral resolution of 3–10 nm [8,10]. The indirect methods refer to the acquisition of derived variables such as the normalized difference vegetation index (NDVI) and the enhanced vegetation index (EVI) from remote sensing data [11]. The variables are then combined with field data to construct models to classify tree species. Light detection and ranging (LiDAR) is a remote sensing technology that can be employed to obtain the three-dimensional geographic coordinates of a research object, especially information relating to vertical structure, directly, quickly and accurately [12]. It can also guarantee the efficient and accurate monitoring of a target from different spatial scales [13]. LiDAR works by recording the transmitted and echoed signals of high-frequency pulses capable of penetrating the vegetation canopy gaps [13]; these signals are then compared and appropriately processed to obtain parameters such as the distance, coordinates, and height of the target [14]. Therefore, LiDAR can be used for studies on tree height [15], canopy density [16], leaf area index [17], individual tree detection [18,19] and tree species classification [20].
From a sensor perspective, there are three common approaches to tree species classification when using LiDAR: using a single LiDAR to extract features such as structure and intensity to classify tree species; using multiple LiDAR to extract structural and intensity features, vegetation index features, etc., to classify tree species; and using LiDAR with other sensors (e.g., multispectral, hyperspectral) to extract structural features, spectral features, vegetation index features, etc., to classify tree species. The first data processing approach is relatively simple, and the potential errors in data caused by sensor differences can be avoided, but the classification accuracy is limited. For example, Ørka et al. (2007) [21] constructed mean intensity features and intensity standard deviation features of ITC based on multi-echo data from a single LiDAR and classified tree species by principal component analysis and linear discriminant analysis with a classification accuracy of 68–74%. Vaughn et al. (2012) [22] used a support vector machine (SVM) to classify five tree species in the Pacific Northwest of the United States based on the structural information and average intensity information provided by discrete point LiDAR data, with an overall accuracy of 79.2% (kappa = 0.74). Hamraza et al. (2019) [23] also used LiDAR intensity information to classify needles and broadleaf trees, and their classification accuracy was 65–90%. The second approach, i.e., based on a combination of multiple LiDAR data, has greater potential for accuracy and quantity of tree species classification but also suffers from relatively cumbersome data collection and processing, in addition to potential errors due to discrepancies in data from different sources. For example, Korpela et al. (2010) [24] realized the classification of three tree species in Finland based on two kinds of LiDAR data, with an accuracy of 88–90%. Vaughn et al. (2012) [22] combined airborne waveform LIDAR with discrete point LIDAR to improve the overall accuracy from 79.2% (kappa = 0.74) to 85.4% (kappa = 0.817) for five species. Brindusaet (2018) [25] constructed intensity features and vegetation index features to classify ten tree species based on intensity information from LiDAR at 1550 nm, LiDAR at 1064 nm, and LiDAR at 532 nm with an accuracy of 57.6–90.6%. The third approach uses LiDAR with other sensors (e.g., multispectral, hyperspectral) to classify tree species [26]. This method can fully obtain vegetation spectral and structural information to classify tree species accurately. However, LiDAR data are often used as secondary information. Combining the homo-spectral and hetero-spectral features of optical remote sensing data limits the accuracy of tree species classification to some extent. For example, Shi et al. (2018) [27] used a random forest classification model to classify 15 common urban tree species by combining structural parameters extracted from LiDAR data and a vegetation index extracted from hyperspectral data, with an overall accuracy of 70%. Kukkonen et al. (2019) used multispectral and airborne LiDAR data to predict three tree species in eastern Finland, with an overall accuracy of 88.2% (kappa = 0.79) [28]. Moreover, hyperspectral sensors are expensive compared with LiDAR sensors, and data processing is more complicated.
In addition to precise 3D coordinates, most LiDAR systems also record “intensity”, loosely defined as the strength of the backscattered echo for each measured point. LIDAR intensity data have proven beneficial in species community classification and richness estimation because they are related to surface parameters such as reflectance [29,30]. However, a literature review shows that most previous studies on the fine classification of tree species based on UAV LiDAR based on limited intensity features and fewer tree species. Korpela et al. (2010) [24] and Heinzel and Koch (2011) [31] derived intensity features such as Imean, Istd, Iskewness and Ikurtosis, which represent the mean, standard deviation, skewness and kurtosis of the intensity values for all of the laser points within a crown. Tomohiro et al. (2017) [32] derived intensity image features, which represent RGB images generated using LiDAR intensity. Brindusaet et al. (2018) [25] derived PE, IR, D1_2, RM, and NDIR, which represent the intensity value at percentiles, the standard deviation of the mean intensity of each ring after dividing the tree point cloud into rings, the difference of intensity between the first and second returns of the same pulse in the same channel, the ratio between different statistics, and the infrared normalized difference vegetation index. Therefore, how to use LiDAR intensity for fine tree species classification is a problem worth studying. To further explore the intensity of data-based tree species classification, we propose the concept of intensity frequency. Intensity frequency is defined as the number of times a certain intensity value appears in the ITC point cloud.
Currently, the single-echo scanning speed for mainstream commercial LiDAR ranges between 100,000 and 500,000 points/s. Reducing the number of laser points emitted per second can reduce the power consumption per unit time and improve the efficiency of data acquisition, but it will affect the intensity frequency of ITC.
In this study, eight major tree species in the study area are used as objects, and 147 intensity frequency features are constructed based on unmanned aerial vehicle laser scanning (UAV) data for tree species classification using a random forest (RF) model. The accuracy of tree species classification at four different point cloud densities is analyzed.
Our overall aim was to gain an understanding of the species discrimination potential of intensity frequency features. The specific research aims were to:
1. Describe the differences of intensity frequency between interspecies and intraspecies;
2. Demonstrate the ability of intensity frequency feature in tree species classification;
3. Quantify the effects of point cloud density in species classification;
4. Examine the consistency between feature selected by MR and feature selected by RF.

2. Study Area and Method

2.1. Study Area

The study area is located on the East Lake campus of Zhejiang A&F University, Lin’an District, Hangzhou City, Zhejiang Province, China, as shown in Figure 1. The East Lake campus covers an area of 1.3 km2. The main tree species include Ginkgo Biloba, Acer Buergerianum Miq, Cinnamomum Camphora, Magnolia Grandiflora, Celtis Sinensis Pers, Glyptostrobuspensilis, Michelia Figo, and Salix Babylonica, which are common green tree species in cities in subtropical regions.

2.2. Field Data

An urban forest is mainly characterized by blocks, strips, and individual trees, with a fragmented distribution and heterogeneous understory, which is very different from a large and continuously distributed forest in the general sense [33,34]. In this study area, there are more trees in a strip distribution, and the understory vegetation is mostly low shrubs and herbaceous plants. Based on the above characteristics, 19 sample strips were established in the study area for ground surveys using a UAV. The ground survey classified 1106 trees and eight species, namely, Ginkgo Biloba (G.B: n = 489), Michelia Figo (M.F: n = 91), Cinnamomum Camphora (C.C: n = 191), Celtis Sinensis Pers (C.S: n = 56), Acer Buergerianum Miq (A.B: n = 73), Salix Babylonica (S.B: n = 91), Glyptostrobuspensilis (G.L: n = 68), and Magnolia grandiflora (M.G: n = 47). Standing trees with a diameter at breast height (DBH) greater than 5 cm were measured in each strip. Measurements included individual tree position, species, height, canopy height, and crown radius due north and south. The DBH of the individual trees was measured using a DBH ruler. We use two real time kinematics (RTKs) of the same model (CHC i70), one as a reference station and one as a mobile station to measure the position of the tree. Treetop height was measured using a Blume–Leiss altimeter. Since UAV observations are carried out from top to bottom, there is a small probability that individual trees in the middle and low layers will be detected by a single-echo UAV [35,36]. The lower and middle vegetation is often not the same species as those studied here, which will directly increase the intensity complexity and make tree species classification more difficult. Therefore, these individual trees were automatically excluded from the data analysis and tree species classification process of this study.

2.3. LiDAR Data Acquisition and Processing

In this study, a DJI Matrice 600 Pro six-rotor UAV was used as the remote sensing platform, equipped with a lightweight Velodyne Puck LITE™ LiDAR sensor and combined with a ground CHC i70 RTK receiver measurement system for data acquisition, as shown in Figure 1d. Due to equipment constraints, the main LiDAR data were obtained on 17 April 2021, and supplements were obtained on 15 May 2021. The tree species G.B, M.F, C.C, A.B, S.B, G.L, and M.G are segmented as ITC using LiDAR data from 17 April; C.S is segmented as ITC using LiDAR data from 15 May. During the flight, the UAV had an average flight height of 60 m, an average flight speed of 8 m/s, and a flight path spacing of approximately 25 m, and the lateral overlap rate of data sampling was approximately 50%. The sensor recorded a laser pulse wavelength of 903 nm, maximum scanning angle of ±15°, scanning frequency of 20 Hz, and scanning speed of 300,000 points/s. Considering that the multiple echoes of LIDAR may cause interference in the intensity frequency, we only selected the first echo information. The average point cloud density was obtained using the total number of point clouds after denoising, divided by the area, which was 230 points/m2. The detailed LiDAR parameters are listed in Table 1.
As shown in Figure 2, the preprocessing of LiDAR data includes the original point cloud decomposition, denoising, separation of ground points, and normalization. The process is as follows: The LiDAR data, GNSS antenna data, and ground CHC i70 RTK data are processed in the accompanying commercial software Inertial Explorer, and then the LAS point cloud is exported in ZtLiDAR V2.2.0 software. In this paper, the point cloud format is LAS1.4. Then, the noises were denoised, including the very high point caused by other low-flying UAVs or birds and the very low point formed by the computational anomalies, and then the denoised point clouds were separated from the ground points and normalized in the LiDAR360 software.

2.4. ITC Information Acquisition

2.4.1. Extraction of ITC

In this study, the point cloud segmentation (PCS) algorithm was used for ITC extraction [37]. It is a top-to-bottom region-growing approach used to segment trees individually and sequentially from the point cloud [8]. The algorithm starts from a tree top and “grows” an individual tree by including nearby points based on the relative spacing. Points with a spacing smaller than a specified threshold were classified as the target tree, and the threshold was approximately equal to the crown radius [8]. Additionally, the shape index (SI) was added to improve segmentation accuracy by avoiding the elongated branch. The PCS algorithm was implemented using LiDAR360 software, and the space threshold was set as the corresponding average crown diameter of each tree species. The segmentation result was a single point cloud cluster tree, and each point cloud in the same cluster tree had the same tree ID. Finally, the crown of each tree was obtained by filtering ground points and low vegetation points through height filtering, and the height filtering threshold was set to 2 m.
The accuracy of segmentation is evaluated by using the following three metrics: recall ( r represents the tree detection rate in Equation (1)), precision (p represents the precision of detected trees in Equation (2)), and F1-score (F1 presents the overall accuracy taking both omission and commission into consideration in Equation (3)).
r = Nt Nt + No  
p = Nt Nt + Nc
F 1 = 2 × r × p r + p
where Nt is the number of objects correctly identified as trees, No is the number of trees omitted by the PCS algorithm, and Nc is the number of objects wrongly identified as trees.

2.4.2. Intensity Correction of ITCs and Resampling

The LiDAR intensity depends on the power of the backscattered laser pulse measured by the sensor [38]. According to the radar equation, during transmission, the LiDAR intensity is mainly affected by the spectral reflectance of the surface ( ρ ), the sensor receiving the emitted laser power ( P E ), the LiDAR optical transmission characteristics ( η s y s ), the distance (R), the angle of incidence ( θ ), the receiver aperture diameter (D), and the atmospheric attenuation ( η a t m ) [39]. To reduce the intensity errors in the transmission process caused by the above factors, we model the relationship between the intensity values and the system variables from the LiDAR equation. For the extended Lambert scatterer, the LiDAR equation can be simplified to Equation (4) [40].
    P r = π P E ρ D 2 ( 4 R 2 ) η a t m η s y s | cos θ |
where ρ is the spectral reflectance of the surface. PE, D, ηsys, and ηatm are assumed to be constant C during the same flight [41]. This leads to Equation (5).
P r = C   ρ R 2 | cos θ |
          C = π P E D 2 16 η a t m η s y s
From Equation (5), the LiDAR equation relates the received power P r to the system variables (R, θ ), so the relationship between the intensity value I and the system variables can be converted into the relationship between the intensity value and the received power P r . Inside the LiDAR receiver, P r is finally converted into a calibrated integer (digital number, DN), and this integer is the intensity value I in raw point cloud data. The logarithmic correction model of the laser intensity value proposed by Tan et al. (2014) [40]. To modify the influence of different incident angles at different distances on the LiDAR intensity the logarithmic correction model is selected. The logarithmic model assumes that the receiver converts the received power logarithmically and then converts it to the laser intensity value, as in Equation (7):
I + τ = K 1 l n p r + K 2
where I is the uncorrected intensity value and K 1   and   K 2 are model coefficients.
The purpose of the intensity correction is to correct the intensity for the variables (R, θ ), so that this value depends only on the target property information ( ρ ); therefore, from Equation (7), P r can be replaced by cos θ / R 2 , and Equation (7) can be written as Equation (8)
I + τ = K 1 ln | cos θ | R 2 + K 3  
where τ is the observation error; K 3 = K 2 + lnC.
After the logarithmic change of Equation (8), the logarithmic correction model of the LiDAR intensity value can be obtained as Equation (9):
I s = ( I K 3 ) ln | cos θ s | ln ( R s 2 ) ln | cos θ | ln R 2 + K 4  
where R s is the reference distance and I s is the corrected intensity value, K 4 = K 3 + τ .
In this study, the reference distance was set as the average flight altitude of the UAV of 60 m, and the reference incident angle was set to 0°. Since 0 has no logarithm, when the incident angle θ equals 90°, cos θ is set to 1 150 . The laser ranging value R is ( f l i g h t   h e i g h t z ) / cos θ with flight height equal to 60;   K 1   and   K 2 are model coefficients obtained by fitting the least square method using I , R, θ .
The number of points in an ITC point cloud varies greatly for the same tree species due to the different ITC shapes. This can affect the intensity frequency. To reduce the effect of the number of points in an ITC point cloud on the intensity frequency, the frequency value associated to the intensity value is multiplied by a weight. This weight is computed using the inverse distance weighting (IDW) according to Equation (10).
φ i = 1 p i i = 1 n 1 p i
where φ i represents the weight of the number of the i-th ITC point cloud, p i represents the number of points in the i-th ITC point cloud, and n is the number of ITCs.
To analyze the influence of canopy intensity frequency on tree species classification accuracy under different point cloud densities, this study set three sampling rates of 80%, 50%, and 30% to resample the D1 data set by the random thinning method [42]; this can be helpful in estimating the most adequate number of features, and thus reducing the complexity and decreasing the computational time. The point density of the three resampled point clouds, D2, D3, and D4 as well as D1 are shown in Table 2.

2.5. Intensity Frequency Feature Calculation and Difference Analysis

The intensity frequency feature process obtained from the ITC point cloud is shown in Figure 3. First, the already-corrected intensity of the ITC point cloud is assessed to obtain the intensity frequencies. Then, an intensity frequency value is assigned to an intensity frequency feature. For example, suppose an ITC point cloud of G.B with intensity values from 3 to 150, where the number of times the point with intensity 3 appears in this ITC point cloud is 10, is said to be I F 3 = 10. After multiplying by φ i and rounding, I F 3 is considered as an intensity frequency feature. This study acquired 147 intensity frequency features from I F 3 to I F 150 .
As shown in Figure 4, the Savitzky-Golay (S-G) filter (Savitzky, 1964) was selected for data stream smoothing [43] to enhance the characteristics of each ITC intensity frequency curve. The S-G filter is a filtering method based on a local polynomial least squares fit. The most important feature of this filter is that the shape and width of the signal can be kept constant while filtering the noise. This process is implemented by calling ‘scipy. signal’ in Python, where the smoothed window length M is taken as 51, the polynomial order is set to 3, and the fill signal of the filter is set to the nearest mode.
To test the hypothesis that intensity frequencies differ significantly among tree species and not significantly among the same tree species, a nonparametric Kruskal-Wallis (KW) test [44] was used, followed by a comparison of the individual tree species intensity frequency features using the mean rank (MR). Post hoc analyses for KW and MR were performed using SPSS (IBM v.26).

2.6. Random Forest and Tree Species Classification

A random forest (RF) is a combination model with many decision trees [45]. The classification process is as follows: bootstrapping is used to form N bootstrap samples, a classification tree model is established for each bootstrap sample, and the sample corresponds to all the training data and test data of the classification tree model. Finally, the model of N classification tree results is taken as the final classification result.
For training sample data with a large number of differences, a sample with a dominant number will have a greater impact on the random forest. The attribute weights produced by an RF on this kind of data are not credible. To maintain a relatively balanced number of tree species, 446 samples were taken in the study to use in the classification of the following tree species: Ginkgo Biloba (G.B: n = 86), Michelia Figo (M.F: n = 52), Cinnamomum camphora (C.C: n = 64), Celtis Sinensis Pers (C.S: n = 27), Acer Buergerianum Miq (A.B: n = 71), Salix Babylonica (S.B: n = 78), Glyptostrobuspensilis (G.L: n = 41), and Magnolia grandiflora (M.G: n = 27). The training and testing data sets were divided into 313 and 133 samples at a ratio of 7:3. The number of decision trees was set to 500 to ensure that each sample was classified more than once. A total of 147 intensity features were input into the RF model. The RF was performed in Python.

2.7. Tree Species Classification Process

In summary, the flow chart of this research is shown in Figure 5. First, the raw data are decoded to LAS format point clouds. Then, the LAS point clouds are denoised, normalized, and segmented to the ITC point clouds. ITC point clouds are labeled as tree species using field data in ArcGIS. Then, intensity correction data set (100%) were resampled as other three density data sets and the 147 intensity frequency features of each ITC of each data set are obtained. Finally, four data sets are classified with RF and then assessed.

2.8. Accuracy Evaluation

The measures of the per-tree species classification include producer accuracy (PA, Equation (11)), user accuracy (UA, Equation (12)), commission errors (CE, Equation (13)), and omission errors (OE, Equation (14)). The measure used for the all-tree species classification is the overall accuracy (OA, Equation (15)).
P A = N r r N r r + N r n
U A = N r r N r r + N n r
C E = N n r N r r + N n r = 1 UA
O E = N r n N r r + N r n = 1 P A
O A = N c o r r N t o t a l = N r r N r r + N n r + N r n
where for a given tree species, N r r represents the number of correctly classified samples, N r n represents the number of samples of that species wrongly classified, and N n r represents the number of samples of another tree species wrongly classified as the given species. The OA is the total number of samples of any tree species classified correctly ( N c o r r ) divided by the total number of validation samples ( N t o t a l ).

3. Results

3.1. ITC Extraction Results

Figure 6a gives an example of G.B point clouds and Figure 6b shows the ITC detected with the PCS algorithm. As seen in Figure 6b, the PCS algorithm is able to detect ITC successfully. A total of 987 trees were correctly detected within the eight tree species using the PCS algorithm, accounting for 89.2% of the total trees surveyed; 123 (11.8%) ITCs were not detected, and the number of ITCs that were not present by incorrect detection was 139 (13.6%). Table 3 shows the ITC detection precision relative to the eight tree species. As can be seen from Table 3, the detection accuracy of ITC is promising, with M.F having the highest detection accuracy of 93.4%, followed by C.S with an accuracy of 91.1%, and G.L with a relatively low detection accuracy of 82.4%.

3.2. Intensity Correction Results

The average intensity frequencies of each tree species extracted based on Equations (4)–(10) are shown in Figure 7. The average intensity frequency of a tree species is the average intensity frequency of all ITC point clouds of this species. Figure 7a shows that the average ITC intensity frequency of the eight species is different, among which C.S is the highest and G.B is the lowest. These differences provide the conditions for the classification of tree species.
Figure 7b shows the standard deviation of each tree species’ average intensity frequency after intensity correction. The mean standard deviation of the intensity of a tree species is the average of the standard deviation of the intensity values of all ITC point clouds for this species. The average standard deviation after correction is reduced relative to before correction, and this is likely to be because of the logarithmic correction model. This model reduces the error in intensity caused by the laser beam range and incident angle, and it produces an intensity value that is mainly a function of the target reflectance characteristics.

3.3. Results of Intensity Frequency of Different Species

Figure 8a shows a diagram of point clouds with sampling rates of 100%, 80%, 50%, and 30% for the eight tree species. As seen in Figure 9a, from top to bottom, the point cloud gradually becomes sparse as the sampling rate decreases, but the canopy outline and key points remain relatively intact, which is expected because the segmentation was carried out with the full density point cloud D1. From left to right, the canopy shape, canopy point cloud imaging, and point cloud density of different tree species vary. For example, the point cloud of G.B is relatively sparse compared with the other seven species, which may result from the fact that G.B itself has smaller leaves and a large gap between leaves and does not form a tightly expanded canopy shape, thus resulting in a smaller number of point clouds. This shows that point clouds can reflect the differences in leaf shape and canopy structures of different trees.
Figure 8b shows the intensity frequency curves of point clouds at different densities of the eight tree species. From left to right, the characteristics of the intensity frequency curve of each tree species are different, such as peak value, wave peak number, and maximum frequency, providing conditions for the classification of tree species using intensity frequency. For a given tree species, the shape of the curve is similar as the sampling rate decreases, but the troughs show an overall decreasing trend. This may be due to the fact that the trough itself is small, and the relative number decreases more after random sampling, so the curve shows a downward trend of the trough. The decrease in sampling rate, i.e., the decrease in point cloud density, causes the curve to lose some details, which may lead to a decrease in the correct classification rate of tree species and may lead to the misclassification of tree species. As an example, when the sampling rate is 30% for C.S, the second wave of its frequency curve disappears. The shape and peak of the intensity frequency curve are similar to those of S.B and G.L, which may lead to C.S being misclassified as S.B and G.L. This further indicates that the ability of point cloud intensity frequencies to represent differences in tree species is influenced by point cloud density.

3.4. Intensity Frequency Difference Analysis Results

The results of the intensity frequency difference analysis are shown in Table 4. Table 4 shows that the p values of the Kruskal–Wallis H-test of the intraspecies intensity frequencies of the eight trees are all greater than 0.05, while the interspecies p values are all less than 0.01, indicating that the intraspecies differences in the intensity frequency are not significant, but the interspecies intensity frequency differences are significant. This provides a theoretical basis for distinguishing eight tree species based on intensity frequency. In addition, Table 4 also shows the impact of intensity frequency sampling on tree species classification differences. The analysis shows that the interspecies statistic H(K) decreases gradually with the decrease in sampling rate, while the intraspecies p and H(K) remain unchanged. This may indicate that the overall statistical variation of interspecies intensity frequencies decreases as the number of points in the point clouds decreases; the statistical variation of intraspecies intensity frequencies remains constant with some stability.
Figure 9 shows the MR histogram of each tree species under 147 features. There are eight bars above each feature, indicating eight tree species. The height of the bars represents the mean rank (MR) of that tree species under this feature. Statistically, if the MR differs significantly, the population to which the two groups of samples belong is considered significantly different. By looking at the bars with significant differences in height, it is possible to initially see which tree species are more different from other species in terms of features. In this study, 4% of the mean MR was used as the threshold [46] to screen the most relevant features of various tree species, as summarized in Table 5.

3.5. Screening Results of Important Random Forest Features

Figure 10 shows the top 10 features with the most relevant importance for tree classification of the eight tree species using the RF model in D1. As seen in the figure, the top 10 most relevant features of the eight tree species accounted for more than 34% of the total 147 features. Among them, the most relevant of the top 10 features of A.B accounted for 88%.
Comparing Figure 10 and Table 5, it can be seen that the priori feature selected by MR covers the top 10 posterior features selected by RF. This can provide a reference for selection of the most relevant features using the MR before classification in the future.
For M.F and G.L, the proportions of the top 10 features are 34% and 36%, respectively, consistent with the classification results in Section 3.6. A.B is stable and has the high UA and PA among all species, indicating that the top 10 features are relevant, the tree species have prominent features, and the classification accuracy is high.

3.6. Tree Species Classification Results

The results of tree species classification with an RF model based on 147 intensity frequency features are shown in Table 6. As seen in Table 6, the UA of tree species classification can range from 100% down to 82.6%. The PAs of three tree species reached more than 90%, three species reached more than 85%, and two species had PAs of 75.0%. The lowest CE is 0%, and the lowest OE is 4.8%. AU is higher than PU, and CE is lower than OE. This shows that intensity frequency can provide effective information for tree species classification. The accuracy of tree species classification using intensity frequency features is high.
As shown in Table 6, Table 7, Table 8 and Table 9, the sum of correct classification trees (SCI) decreases as the sampling rate decreases: SCI equals 117 at D1 data set, 97 in D2, 74 in D3, and 65 in D4. The largest decrease in SCI occurs between the D2 and D3. This is probably because the sampling rate decreases the most (by 30%) from D2 to D3. The number of point clouds is reduced more, and more key points that reflect the differences in intensity frequencies of each tree species may be lost, leading to a decrease in classification accuracy. Therefore, the SCI at an 80% sampling rate differs the most from 50% relative to the other two cases, indicating that the point cloud density positively correlates with the sum of correct classification trees (SCIs) in this study.
UA remains relatively stable across the four sampling rates in the data set. The reason may be because most of the classified samples after random sampling still provide the critical information required by RF for successful classification. The UA remains relatively stable, which is consistent with the implication shown in Figure 8 that the intensity frequency curve retains its basic shape after resampling, but the details change. This indicates that the intensity frequency feature can maintain a relatively robust UA at different point cloud densities in this study. The PA decreases with decreasing sampling rate, potentially because as the sampling rate decreases, some of the intensity frequency features can no longer provide the information needed for RF to classify tree species, leading to an increase in the number of true samples classified as other ( N r n ).
Finally, the overall accuracy (OA) is above 84% for all four data sets, with the highest being the D2 data set (OA = 87.4%), slightly higher than that of the D1 data set. The reason could be that as the sampling rate decreases, some intensity frequency features of ITC, although able to be classified by RF at a 100% sampling rate, have a high error rate. RF can no longer classify them at an 80% sampling rate, which leads to an increase in N r n since the overall accuracy = ( N r r )/( N r r + N n r + N r n ). While the N r r at an 80% sampling rate decreases, the N r n also decreases. The numerator decreases less than the denominator, resulting in a slight increase in overall precision.
In summary, the intensity frequency features of four data set have promising results on UA, PA, CE, OE, OA, and SCI. This indicates that the intensity frequency features of the ITC point cloud of eight tree species obtained from this study can accurately perform tree species classification.

4. Discussion

The results show that the LiDAR intensity frequency, combined with the PCS and RF algorithms, can achieve high-precision classification of eight urban forest tree species in the study area.
Previous studies have shown that intensity data can realize surface classification and object detection [30,47]. The target water content significantly influences the intensity [48,49]. Both the reflectivity and the structural characteristics of ITC would influence the intensity of the intensity of point clouds acquired by UAV-LiDAR [49]. Different tree species may have different water contents, canopy structures, and physiological and biochemical characteristics of leaves [50]. Therefore, the intensity features of LiDAR have promising potential for tree species classification.
LiDAR intensity data have some advantages for tree species classification over passive remote sensing sensor data such as multispectral and hyperspectral samples [25,51,52]. First, the measurement of LiDAR intensity has the advantage of being independent of external lighting conditions [31], so LiDAR data are not affected by variable shadows and natural light, and the positional accuracy of point clouds is, in general, more reliable when compared to that of point clouds produced with aerial images [53]. Second, although passive remote sensing imaging data can be reversed to generate point clouds, in vegetated areas, such point clouds have limited penetration and ground points. The LiDAR point cloud will contain background information, contrary to that produced with images, allowing to the structuring of information. In addition, the incidence angle and height information provided by LiDAR is brought into some intensity correction models. This can reduce the intensity data differences caused by distance and angle of incidence to some extent so that the intensity contains the difference in target reflectivity [30]. Finally, LiDAR can provide both structural and intensity information, which has the advantages of requiring less storage space, and less complex and time-consuming data processing compared to hyperspectral and multispectral data [54,55]. There are also advantages when using intensity frequencies versus individual intensities for tree species classification. A single intensity can only reflect the characteristics of one point of the whole tree canopy and carries limited information. If there is a point cloud of reflections from other features, it may cause significant interference with the classification results. In contrast, intensity frequency is the collection of all intensities of the whole canopy, i.e., carrying intensity information and having a small degree of anti-interference ability. The intensity frequencies can provide pseudo-waveform features similar to the spectral reflectance curves. The intensity frequency feature of pseudo-waveform could provide more tree species differences information than single-valued features.
Current studies using intensity for tree species classification include converting intensity into raster images and then combining other features for species classification [23,47,56,57] or extracting the mean and standard deviation of intensity from manually depicted tree canopies for species classification [58]. The advantage of this approach is in preserving the spatial distribution characteristics of the intensity data, while the disadvantage is that the unit raster intensity value is the sum of the intensity of the points that are projected into the pixel or the standard deviation of those intensity values, which may somewhat lose the structure of the target object to which the intensity responds. The other method involves constructing vegetation indices for tree species classification using multiple LiDAR sensors with different wavelengths; for example, Yu et al. (2017) classified pine, spruce, and birch trees based on point cloud single-channel intensity, multichannel intensity, and combined data of three channels with a maximum accuracy of 85.9% [25]. The advantage of this method is that it enhances the differences between different tree species by constructing vegetation indices, thus making it easier to classify tree species. The disadvantage is that acquiring data from different channels requires multiple repetitive operations with multiple sensors, which is time-consuming and expensive. Additionally, the vegetation indices that can induce the response of the intensity curves of each tree species are not yet clear, which is also a worthwhile research direction in the future.
Although we obtained a qualitative result showing that the intensity frequencies of the eight tree species did not differ significantly for intraspecies or interspecies observations by the KW test, we did not perform a more detailed quantitative analysis for each tree species. Future research could be conducted for a quantitative analysis of the differences in intensity frequencies among tree species.
In Table 4, the PU and UA of some species need further improvement, such as the PA of both G.L and M.G being 75%. The main reason for the low PA accuracy of these two species may be that the training samples of these two species are few, and the RF model cannot thoroughly learn the intensity frequency characteristics of these two species. As a result, the intensity frequency does not possess the tree’s complete canopy characteristics, resulting in a decrease in the PA. At the same time, the classification error of tree species in this study may also be derived from two aspects. First, the difference in canopy structure information and the crown shapes of different individual plants of the same tree species are often similar but different. Second, the physiological and biochemical status of target leaves may also bring about tree species classification errors. Different crown properties affect the distribution of laser echoes within and on the surface of the tree crowns [59], thus affecting the intensity frequency. More species lead to more diversity in leaf shape, size, and reflectance, as well as in tree structure, resulting in more difficult classification tasks. In this case, up to eight species are classified using only the intensity signatures from a near-ground flight (60 m) UAV, with an overall accuracy of over 86%. To our knowledge, such results have not been reported in the scientific literature.
This study was based on ITC point clouds for point cloud thinning and analyzed the effect of density on the classification accuracy of tree species. This approach followed the principle of controlling a single variable for the comparison test [60]. Density as the only dependent variable when studying the effect of density on classification results. The effect of canopy segmentation differences on classification accuracy was excluded, because point clouds with different densities would lead to different results of PCS segmentation of canopies. Therefore, the conclusion related to the influence of point cloud density on classification accuracy in this study does not apply to point clouds of different densities acquired at different heights. This is also a direction worth investigating. There are some limitations in this study. First is the limitation of ITC segmentation; the method of this study is applicable to point clouds so that ITC can be segmented accurately. However, in complex terrain, complex forest conditions, and an automatic acquisition of segmented canopy, unavoidable random or systematic errors may arise, which leads to elevated classification errors [61]. The second limitation is that the topography of this study area is relatively flat. If the topography is complex, the data noise of each canopy will be greater, and the classification will be less reliable. In addition, during intensity correction, we treat the observation error τ as a constant, which may also cause errors in the intensity correction. Finally, there is the diversity of tree canopies. Theoretically, different tree species have different spectral reflectance properties, and the same tree species has similar spectral reflectance properties [62]. However, the diversity of canopies caused by the variation in canopy structure, leaf shape, and reflectance with age increases the intraspecific variability and classification error probability [25]. Therefore, further research is necessary to follow up on whether stratification by tree height or age can improve classification performance. In addition, random forest models, trained using a limited number of samples from a specific region, are difficult to apply on a broader scale without additional training samples to account for site variation.

5. Conclusions

In this study, we propose a tree species classification method based on UAV-LiDAR intensity frequency features, i.e., extracting intensity frequency features from LiDAR data and combining the PCS algorithm to segment the acquired point cloud into ITC and the random forest model to classify the main tree species in the study area. In the research process, we extracted the intensity frequency features of different tree species from resampled ITC point clouds of four densities, and then used the RF model to classify the tree species. The research results show that the intensity frequency features of the point clouds of the four densities can achieve higher accuracy in classifying the eight major tree species in the study area, and the overall accuracy of tree species classification can reach more than 84% even if the point cloud density is resampled to 30%. Therefore, the UAV-LiDAR intensity frequency feature can be used for the classification of different tree species. In addition, intensity frequency feature selection is a relevant guarantee to achieve tree species classification accuracy, and mean rank (MR) is a promising method to achieve important intensity frequency feature selection.
In our future work, we will carry out further research in the intelligent segmentation of individual tree crown, and LiDAR intensity frequency combined with hyperspectral remote sensing data to achieve intelligent identification of tree species, which is another hot spot in the current related research.

Author Contributions

Project administration, H.D. and X.L.; funding acquisition, H.D. and X.L.; resources, G.Z.; supervision, F.M. and X.L.; investigation, X.L., L.Z., B.Z., J.X. and D.Z.; data curation, B.Z.; conceptualization, Y.G. and H.D.; methodology, Y.G.; software, Y.G.; validation, Y.G.; formal analysis, Y.G.; writing—original draft preparation, Y.G.; writing—review and editing, Y.G. and H.D.; visualization, Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

The authors gratefully acknowledge the support of the National Natural Science Foundation of China (U1809208, 32171785, 32201553), Leading Goose Project of Science Technology Department of Zhejiang Province (2023C02035).

Acknowledgments

The authors gratefully acknowledge the supports of various foundations. The authors are grateful to the editor and anonymous reviewers whose comments have contributed to improving the quality of this study.

Conflicts of Interest

The authors declare that they have no competing interests.

References

  1. Mikael, K.; Lauri, K.; Matti, M.; Aki, S.; Petteri, P. How much can airborne laser scanning based forest inventory by tree species benefit from auxiliary optical data? Int. J. Appl. Earth Obs. Geoinf. 2018, 72, 91–98. [Google Scholar] [CrossRef]
  2. Gogoi, A.; Ahirwal, J.; Sahoo, U. Evaluation of ecosystem carbon storage in major forest types of Eastern Himalaya: Implications for carbon sink management. J. Environ. Manag. 2022, 302, 113972. [Google Scholar] [CrossRef] [PubMed]
  3. Litza, K.; Alignier, A.; Closset-Kopp, D.; Ernoult, A.; Mony, C.; Osthaus, M.; Staley, J.; Berge, S.V.D.; Vanneste, T.; Diekmann, M. Hedgerows as a habitat for forest plant species in the agricultural landscape of Europe. Agric. Ecosyst. Environ. 2022, 326, 107809. [Google Scholar] [CrossRef]
  4. Pisarek, P.; Bueno, M.; Thiry, Y.; Legout, A.; Gallard, H.; Le Hécho, I. Influence of tree species on selenium and iodine partitioning in an experimental forest ecosystem. Sci. Total Environ. 2022, 809, 151174. [Google Scholar] [CrossRef]
  5. Cavender-Bares, J.; Schneider, F.D.; Santos, M.J.; Armstrong, A.; Carnaval, A.; Dahlin, K.M.; Fatoyinbo, L.; Hurtt, G.C.; Schimel, D.; Townsend, P.A.; et al. Integrating remote sensing with ecology and evolution to advance biodiversity conservation. Nat. Ecol. Evol. 2022, 6, 506–519. [Google Scholar] [CrossRef]
  6. Neyns, R.; Canters, F. Mapping of Urban Vegetation with High-Resolution Remote Sensing: A Review. Remote Sens. 2022, 14, 1031. [Google Scholar] [CrossRef]
  7. Fassnacht, F.E.; Latifi, H.; Stereńczak, K.; Modzelewska, A.; Lefsky, M.; Waser, L.T.; Straub, C.; Ghosh, A. Review of studies on tree species classification from remotely sensed data. Remote Sens. Environ. 2016, 186, 64–87. Available online: https://sci-hub.se/10.1016/j.rse.2016.08.013 (accessed on 1 January 2021).
  8. Shen, X.; Cao, L. Tree-Species Classification in Subtropical Forests Using Airborne Hyperspectral and LiDAR Data. Remote Sens. 2017, 9, 1180. [Google Scholar] [CrossRef] [Green Version]
  9. García-Pardo, K.A.; Moreno-Rangel, D.; Domínguez-Amarillo, S.; García-Chávez, J.R. Remote sensing for the assessment of ecosystem services provided by urban vegetation: A review of the methods applied. Urban For. Urban Green. 2022, 74, 127636. [Google Scholar] [CrossRef]
  10. Qin, H.; Zhou, W.; Yao, Y.; Wang, W. Individual tree segmentation and tree species classification in subtropical broadleaf forests using UAV-based LiDAR, hyperspectral, and ultrahigh-resolution RGB data. Remote Sens. Environ. 2022, 280, 113143. [Google Scholar] [CrossRef]
  11. Hovi, A.; Korhonen, L.; Vauhkonen, J.; Korpela, I. LiDAR waveform features for tree species classification and their sensitivity to tree- and acquisition related parameters. Remote Sens. Environ. 2016, 173, 224–237. [Google Scholar] [CrossRef]
  12. Beaudoin, A.; Hall, R.J.; Castilla, G.; Filiatrault, M.; Villemaire, P.; Skakun, R.; Guindon, L. Improved k-NN Mapping of Forest Attributes in Northern Canada Using Spaceborne L-Band SAR, Multispectral and LiDAR Data. Remote Sens. 2022, 14, 1181. [Google Scholar] [CrossRef]
  13. Rahman, F.; Onoda, Y.; Kitajima, K. Forest canopy height variation in relation to topography and forest types in central Japan with LiDAR. For. Ecol. Manag. 2022, 503, 119792. [Google Scholar] [CrossRef]
  14. Yin, D.; Wang, L. Individual mangrove tree measurement using UAV-based LiDAR data: Possibilities and challenges. Remote Sens. Environ. 2019, 223, 34–49. [Google Scholar] [CrossRef]
  15. Blair, J.; Rabine, D.L.; A Hofton, M. The Laser Vegetation Imaging Sensor: A medium-altitude, digitisation-only, airborne laser altimeter for mapping vegetation and topography. ISPRS J. Photogramm. Remote Sens. 1999, 54, 130–137. [Google Scholar] [CrossRef]
  16. Wortley, L.; Hero, J.-M.; Howes, M. Evaluating ecological restoration success: A review of the literature. Restor. Ecol. 2013, 21, 537–543. [Google Scholar] [CrossRef]
  17. Wang, Y.; Fang, H. Estimation of LAI with the LiDAR technology: A review. Remote Sens. 2020, 12, 3457. [Google Scholar] [CrossRef]
  18. Hershey, J.L.; McDill, M.E.; Miller, D.A.; Holderman, B.; Michael, J.H. A Voxel-Based Individual Tree Stem Detection Method Using Airborne LiDAR in Mature Northeastern U.S. Forests. Remote Sens. 2022, 14, 806. [Google Scholar] [CrossRef]
  19. Corte, A.P.D.; Neto, E.M.D.C.; Rex, F.E.; Souza, D.; Behling, A.; Mohan, M.; Sanquetta, M.N.I.; Silva, C.A.; Klauberg, C.; Sanquetta, C.R.; et al. High-density UAV-LiDAR in an integrated crop-livestock-forest system: Sampling forest inventory or forest inventory based on individual tree detection (ITD). Drones 2022, 6, 48. [Google Scholar] [CrossRef]
  20. Shi, Y.; Wang, T.; Skidmore, A.K.; Heurich, M. Important LiDAR metrics for discriminating forest tree species in Central Europe. ISPRS J. Photogramm. Remote Sens. 2018, 137, 163–174. [Google Scholar] [CrossRef]
  21. Ørka, H.; Naesset, E.; Bollandsås, O. Utilizing Airborne Laser Intensity for Tree Species Classification. In Proceedings of the ISPRS Workshop Laser Scanning 2007 SilviLaser, Espoo, Finland, 12–14 September 2007; Available online: https://www.isprs.org/proceedings/XXXVI/3-W52/final_papers/Oerka_2007.pdf (accessed on 1 January 2021).
  22. Vaughn, N.R.; Moskal, L.M.; Turnblom, E.C. Tree Species Detection Accuracies Using Discrete Point Lidar and Airborne Waveform Lidar. Remote Sens. 2012, 4, 377–403. [Google Scholar] [CrossRef]
  23. Hamraz, H.; Jacobs, N.B.; Contreras, M.A.; Clark, C.H. Deep learning for conifer/deciduous classification of airborne LiDAR 3D point clouds representing individual trees. ISPRS J. Photogramm. Remote Sens. 2019, 158, 219–230. [Google Scholar] [CrossRef] [Green Version]
  24. Korpela, I.; Ørka, H.; Maltamo, M.; Tokola, T.; Hyyppä, J. Tree species classification using airborne LiDAR—Effects of stand and tree parameters, downsizing of training set, intensity normalization, and sensor type. Silva Fenn. 2010, 44, 319–339. [Google Scholar] [CrossRef] [Green Version]
  25. Budei, B.C.; St-Onge, B.; Hopkinson, C.; Audet, F.-A. Identifying the genus or species of individual trees using a three-wavelength airborne lidar system. Remote Sens. Environ. 2018, 204, 632–647. [Google Scholar] [CrossRef]
  26. Voss, M.; Sugumaran, R. Seasonal Effect on Tree Species Classification in an Urban Environment Using Hyperspectral Data, LiDAR, and an Object- Oriented Approach. Sensors 2008, 8, 3020–3036. [Google Scholar] [CrossRef] [Green Version]
  27. Shi, Y.; Skidmore, A.K.; Wang, T.; Holzwarth, S.; Heiden, U.; Pinnel, N.; Zhu, X.; Heurich, M. Tree species classification using plant functional traits from LiDAR and hyperspectral data. Int. J. Appl. Earth Obs. Geoinf. 2018, 73, 207–219. [Google Scholar] [CrossRef]
  28. Kukkonen, M.; Maltamo, M.; Korhonen, L.; Packalen, P. Multispectral airborne LiDAR data in the prediction of boreal tree species composition. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3462–3471. [Google Scholar] [CrossRef]
  29. Ioki, K.; Tsuyuki, S.; Hirata, Y.; Phua, M.-H.; Wong, W.V.C.; Ling, Z.-Y.; Johari, S.A.; Korom, A.; James, D.; Saito, H.; et al. Evaluation of the similarity in tree community composition in a tropical rainforest using airborne LiDAR data. Remote Sens. Environ. 2016, 173, 304–313. [Google Scholar] [CrossRef]
  30. Kashani, A.G.; Olsen, M.J.; Parrish, C.E.; Wilson, N. A Review of LIDAR Radiometric Processing: From Ad Hoc Intensity Correction to Rigorous Radiometric Calibration. Sensors 2015, 15, 28099–28128. [Google Scholar] [CrossRef] [Green Version]
  31. Heinzel, J.; Koch, B. Exploring full-waveform LiDAR parameters for tree species classification. Int. J. Appl. Earth Obs. Geoinf. 2011, 13, 152–160. [Google Scholar] [CrossRef]
  32. Mizoguchi, T.; Ishii, A.; Nakamura, H.; Inoue, T.; Takamatsu, H. Lidar-based individual tree species classification using convolutional neural network. SPIE 2017, 10332, 193–199. [Google Scholar] [CrossRef]
  33. McBride, J.; Jacobs, D. Urban forest development: A case study, Menlo park, California. Urban Ecol. 1976, 2, 1–14. [Google Scholar] [CrossRef]
  34. Rowntree, R.A. Ecology of the urban forest—Introduction to part II. Urban Ecol. 1986, 9, 229–243. [Google Scholar] [CrossRef]
  35. Avellar, G.S.C.; Pereira, G.A.S.; Pimenta, L.C.D.A.; Iscold, P. Multi-UAV routing for area coverage and remote sensing with minimum time. Sensors 2015, 15, 27783–27803. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Means, J.; Means, J. Use of Large-Footprint Scanning Airborne Lidar To Estimate Forest Stand Characteristics in the Western Cascades of Oregon. Remote Sens. Environ. 1999, 67, 298–308. [Google Scholar] [CrossRef]
  37. Li, W.; Guo, Q.; Jakubowski, M.K.; Kelly, M. A new method for segmenting individual trees from the lidar point cloud. Photogramm. Eng. Remote Sens. 2012, 78, 75–84. [Google Scholar] [CrossRef] [Green Version]
  38. Lohani, B.; Ghosh, S. Airborne LiDAR technology: A review of data collection and processing systems. Proc. Natl. Acad. Sci. India Sect. A Phys. Sci. 2017, 87, 567–579. [Google Scholar] [CrossRef]
  39. Hobbs, S.; Jelalian, A.V. Laser radar systems: Artech house, boston, £64.00 (hb). J. Atmos. Terr. Phys. 1992, 54, 1646. [Google Scholar] [CrossRef]
  40. Tan, K.; Cheng, X.J. Correction of methods of laser intensity and accuracy of point cloud classification. J. Tongji Univ. Nat. Sci. 2014, 42, 131–135. [Google Scholar]
  41. You, H.; Wang, T.; Skidmore, A.K.; Xing, Y. Quantifying the effects of Normalisation of airborne LiDAR intensity on coniferous forest leaf area index estimations. Remote Sens. 2017, 9, 163. [Google Scholar] [CrossRef] [Green Version]
  42. Zhugeng, D.; Lingxiao, W.; Xueliang, J. Effect of point cloud density on forest remote sensing retrieval index extraction based on unmanned aerial vehicle lidar data. Geomat. Inf. Sci. Wuhan Univ. 2021, 41, 711–721. [Google Scholar] [CrossRef]
  43. Savitzky, A.; Golay, M.J.E. Smoothing and differentiation of data by simplified least squares procedures. Anal. Chem. 1964, 36, 1627–1639. [Google Scholar] [CrossRef]
  44. Li, X.; Du, H.; Mao, F.; Zhou, G.; Xing, L.; Liu, T.; Han, N.; Liu, E.; Ge, H.; Liu, Y.; et al. Mapping spatiotemporal decisions for sustainable productivity of bamboo forest land. Land Degrad. Dev. 2020, 31, 939–958. [Google Scholar] [CrossRef]
  45. Rodriguez-Galiano, V.F.; Ghimire, B.; Rogan, J.; Chica-Olmo, M.; Rigol-Sanchez, J.P. An assessment of the effectiveness of a random forest classifier for land-cover classification. ISPRS J. Photogramm. Remote Sens. 2012, 67, 93–104. [Google Scholar] [CrossRef]
  46. Lawrence, R.L.; Wood, S.D.; Sheley, R.L. Mapping invasive plants using hyperspectral imagery and Breiman Cutler classifications (randomForest). Remote Sens. Environ. 2006, 100, 356–362. [Google Scholar] [CrossRef]
  47. Yan, W.Y.; Shaker, A. Radiometric correction and normalization of airborne LiDAR intensity data for improving land-cover classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7658–7673. [Google Scholar] [CrossRef]
  48. Korpela, I.; Koskinen, M.; Vasander, H.; Holopainen, M.; Minkkinen, K. Airborne small-footprint discrete-return LiDAR data in the assessment of boreal mire surface patterns, vegetation, and habitats. For. Ecol. Manag. 2009, 258, 1549–1566. [Google Scholar] [CrossRef]
  49. Yan, W.Y.; Shaker, A. Airborne LiDAR intensity banding: Cause and solution. ISPRS J. Photogramm. Remote Sens. 2018, 142, 301–310. [Google Scholar] [CrossRef]
  50. Michałowska, M.; Rapiński, J. A review of tree species classification based on airborne LiDAR data and applied classifiers. Remote Sens. 2021, 13, 353. [Google Scholar] [CrossRef]
  51. Lefsky, M.; Cohen, W.; Acker, S.; Parker, G.; Spies, T.; Harding, D. Lidar remote sensing of the canopy structure and biophysical properties of douglas-fir western hemlock forests. Remote Sens. Environ. 1999, 70, 339–361. [Google Scholar] [CrossRef]
  52. Ballanti, L.; Blesius, L.; Hines, E.; Kruse, B. Tree Species Classification Using Hyperspectral Imagery: A Comparison of Two Classifiers. Remote Sens. 2016, 8, 445. [Google Scholar] [CrossRef] [Green Version]
  53. Coops, N.C.; Tompalski, P.; Goodbody, T.R.; Queinnec, M.; Luther, J.E.; Bolton, D.K.; White, J.C.; Wulder, M.A.; van Lier, O.R.; Hermosilla, T. Modelling lidar-derived estimates of forest attributes over space and time: A review of approaches and future trends. Remote Sens. Environ. 2021, 260, 112477. [Google Scholar] [CrossRef]
  54. Heinzel, J.; Koch, B. Investigating multiple data sources for tree species classification in temperate forest and use for single tree delineation. Int. J. Appl. Earth Obs. Geoinf. 2012, 18, 101–110. [Google Scholar] [CrossRef]
  55. Dalponte, M.; Bruzzone, L.; Gianelle, D. Tree species classification in the Southern Alps based on the fusion of very high geometrical resolution multispectral/hyperspectral images and LiDAR data. Remote Sens. Environ. 2012, 123, 258–270. [Google Scholar] [CrossRef]
  56. Lang, M.W.; Kim, V.; McCarty, G.W.; Li, X.; Yeo, I.-Y.; Huang, C.; Du, L. Improved detection of inundation below the forest canopy using normalized LiDAR intensity data. Remote Sens. 2020, 12, 707. [Google Scholar] [CrossRef] [Green Version]
  57. Liu, L.; Coops, N.C.; Aven, N.W.; Pang, Y. Mapping urban tree species using integrated airborne hyperspectral and LiDAR remote sensing data. Remote Sens. Environ. 2017, 200, 170–182. [Google Scholar] [CrossRef]
  58. St-Onge, B.; Audet, F.-A.; Bégin, J. Characterizing the height structure and composition of a boreal forest using an individual tree crown approach applied to photogrammetric point clouds. Forests 2015, 6, 3899–3922. [Google Scholar] [CrossRef]
  59. Ørka, H.O.; Næsset, E.; Bollandsås, O.M. Classifying species of individual trees by intensity and structure features derived from airborne laser scanner data. Remote Sens. Environ. 2009, 113, 1163–1174. [Google Scholar] [CrossRef]
  60. Tehseen, S.; Ramayah, T.; Sajilan, S. Testing and Controlling for Common Method Variance: A Review of Available Methods. J. Manag. Sci. 2017, 4, 142–168. [Google Scholar] [CrossRef] [Green Version]
  61. Dalponte, M.; Ørka, H.O.; Ene, L.T.; Gobakken, T.; Næsset, E. Tree crown delineation and tree species classification in boreal forests using hyperspectral and ALS data. Remote Sens. Environ. 2014, 140, 306–317. [Google Scholar] [CrossRef]
  62. Luo, S.; Chen, J.M.; Wang, C.; Gonsamo, A.; Xi, X.; Lin, Y.; Qian, M.; Peng, D.; Nie, S.; Qin, H. Comparative performances of airborne lidar height and intensity data for leaf area index estimation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 300–310. [Google Scholar] [CrossRef]
Figure 1. (a,b) show the geographical location of the study area. (c) is the distribution of the eight tree species. (d) is the diagram of LiDAR data acquisition, and (e) is a schematic diagram of the LiDAR slices of G.B.
Figure 1. (a,b) show the geographical location of the study area. (c) is the distribution of the eight tree species. (d) is the diagram of LiDAR data acquisition, and (e) is a schematic diagram of the LiDAR slices of G.B.
Remotesensing 15 00110 g001
Figure 2. (a) Slice of the point cloud after decoding the raw data; (b) slice of the point cloud after denoising; (c) slice of the separated ground point cloud; and (d) slice of the normalized point cloud.
Figure 2. (a) Slice of the point cloud after decoding the raw data; (b) slice of the point cloud after denoising; (c) slice of the separated ground point cloud; and (d) slice of the normalized point cloud.
Remotesensing 15 00110 g002
Figure 3. Intensity frequency feature acquisition schematic.
Figure 3. Intensity frequency feature acquisition schematic.
Remotesensing 15 00110 g003
Figure 4. A before-and-after diagram of the Savitzky-Golay (S-G) filter; the numbers in the legend are the serial numbers of the 16 ITC point clouds with the tree species G.B; the line colors indicate the intensity frequencies of the different number of ITC point clouds.
Figure 4. A before-and-after diagram of the Savitzky-Golay (S-G) filter; the numbers in the legend are the serial numbers of the 16 ITC point clouds with the tree species G.B; the line colors indicate the intensity frequencies of the different number of ITC point clouds.
Remotesensing 15 00110 g004
Figure 5. Flow chart.
Figure 5. Flow chart.
Remotesensing 15 00110 g005
Figure 6. (a) Point clouds of G.B numbered 1–10; (b) segmentation results using the PCS algorithm (different colors correspond to different ITCs).
Figure 6. (a) Point clouds of G.B numbered 1–10; (b) segmentation results using the PCS algorithm (different colors correspond to different ITCs).
Remotesensing 15 00110 g006
Figure 7. (a) Average intensity frequency of each species; (b) mean standard deviation of the intensity frequency of each species.
Figure 7. (a) Average intensity frequency of each species; (b) mean standard deviation of the intensity frequency of each species.
Remotesensing 15 00110 g007
Figure 8. (a) Pointclouds of different densities of the eight tree species; the vertical coordinate of (b) is the maximum intensity frequency value among all the detected ITC per species. The 8 different colored bars above the same IF feature indicate the MR of the 8 tree species.
Figure 8. (a) Pointclouds of different densities of the eight tree species; the vertical coordinate of (b) is the maximum intensity frequency value among all the detected ITC per species. The 8 different colored bars above the same IF feature indicate the MR of the 8 tree species.
Remotesensing 15 00110 g008
Figure 9. (a,b) Mean rank (MR) histogram of 147 intensity features of 8 tree species. Different tree species are indicated by different colors. the values of MR represent the differences in IF features among the eight tree species.
Figure 9. (a,b) Mean rank (MR) histogram of 147 intensity features of 8 tree species. Different tree species are indicated by different colors. the values of MR represent the differences in IF features among the eight tree species.
Remotesensing 15 00110 g009
Figure 10. Important intensity frequency features for each species selected by RF in D1. The subfigure shows the most relevant features of a tree species. ‘SUM’ equals the sum of the importance of the most relevant features calculated by RF.
Figure 10. Important intensity frequency features for each species selected by RF in D1. The subfigure shows the most relevant features of a tree species. ‘SUM’ equals the sum of the importance of the most relevant features calculated by RF.
Remotesensing 15 00110 g010
Table 1. Summary of remote sensing data acquisition information.
Table 1. Summary of remote sensing data acquisition information.
ParametersInformationParametersInformation
SensorVelodyne Puck LITE ™Ranging Accuracy3 cm
Date of Acquisition2021.4.17, 2021.5.15Mean Point Density230 points/m²
Height60 mWavelength903 nm
Table 2. Density of point clouds.
Table 2. Density of point clouds.
Data Set NameSampling RateMean Point Density
D1100%230 points/m2
D280%184 points/m2
D350%115 points/m2
D430%69 Points/m2
Table 3. ITC segmentation accuracy assessment. The darker color indicates higher accuracy.
Table 3. ITC segmentation accuracy assessment. The darker color indicates higher accuracy.
Tree SpeciesNtNoNcR (%)P (%)F (%)
M.F8561093.4%89.5%91.4%
C.C167241987.4%89.8%88.6%
C.S515791.1%87.9%89.5%
S.B8110889.0%91.0%90.0%
A.B62111384.9%82.7%83.8%
G.B435546789.0%86.7%87.8%
M.G3981183.0%78.0%80.4%
G.L56121582.4%78.9%80.6%
Table 4. Kruskal-Wallis H test results of intraspecies and interspecies.
Table 4. Kruskal-Wallis H test results of intraspecies and interspecies.
Sampling Rate100%80%50%30%
pH(k)pH(k)pH(k)pH(k)
IntraspeciesM.F0.474510.474510.474510.47451
C.C0.476630.476630.476630.47663
C.S0.462250.462250.462250.46225
A.B0.478700.478700.478700.47870
S.B0.479770.479770.479770.47977
G.L0.470400.470400.470400.47040
M.G0.463260.463260.463260.46326
G.B0.480850.480850.480850.48085
Interspecies <0.01181.590<0.01133.241<0.01106.956<0.0187.638
Table 5. Most relevant intensity frequency features for each species selected by MR in D1. ‘ IF a IF b ’ means that all from IF a to IF b were selected.
Table 5. Most relevant intensity frequency features for each species selected by MR in D1. ‘ IF a IF b ’ means that all from IF a to IF b were selected.
Tree SpeciesImportant Intensity Frequency Features in D1
M.F IF 3 , IF 8 - IF 13 , IF 18 , IF 21 , IF 45 , IF 41 - IF 64 , IF 66 - IF 115 , IF 117 - IF 121 , IF 132 , IF 133 , IF 140 , IF 142 - IF 144
C.C IF 4 , IF 6 , IF 8 - IF 16 , IF 44 , IF 49 , IF 50 , IF 55 , IF 60 , IF 63 , IF 64 , IF 70 , IF 73 , IF 74 , IF 76 , IF 101 , IF 106 , IF 110 ,   IF 113 , IF 116 , IF 118 - IF 131
C.S IF 12 - IF 15 , IF 18 - IF 53 , IF 55 - IF 57 , IF 59 , IF 61 - IF 63 , IF 65 - IF 68 , IF 70 - IF 84 , IF 95 , IF 97 - IF 107 , IF 109 ,
IF 112 , IF 114 , IF 116 , IF 119 - IF 121
A.B IF 3 - IF 5 , IF 7 , IF 9 - IF 19 , IF 25 , IF 27 , IF 29 - IF 32 , IF 36 , IF 43 - IF 118 , IF 120 - IF 125 , IF 127 , IF 129
S.B IF 4 , IF 6 , IF 8 - IF 10 , IF 15 , IF 18 - IF 20 , IF 22 - IF 27 , IF 30 , IF 35 , IF 39 , IF 49 , IF 54 , IF 57 , IF 64 , IF 66 , IF 70 , IF 104 ,
IF 105 , IF 109 , IF 113 - IF 116 , IF 120 - IF 133
G.L IF 11 - IF 14 , IF 16 ,   IF 24 , IF 25 , IF 27 , IF 29 , IF 34 , IF 44 , IF 47 , IF 54 , IF 60 , IF 66 , IF 76 , IF 78 , IF 97 - IF 99 , IF 103 , IF 106 ,
IF 109 , IF 111 , IF 119 , IF 121 - IF 123 , IF 131 , IF 134 , IF 135 , IF 143 , IF 145
M.G IF 4 , IF 5 , IF 9 , IF 11 - IF 24 , IF 28 , IF 30 , IF 34 , IF 37 , IF 39 , IF 46 , IF 49 , IF 55 , IF 57 , IF 60 , IF 62 - IF 86 , IF 88 , IF 89 ,
IF 91 - IF 96 , IF 100 - IF 103 , IF 106 - IF 109 , IF 113 - IF 121 , IF 126 , IF 129 , IF 137 , IF 140 , IF 142 , IF 145 , IF 146
G.B IF 4 , IF 6 , IF 9 , IF 11 - IF 13 , IF 19 , IF 21 - IF 23 , IF 25 - IF 30 , IF 120 - IF 125 , IF 146
Table 6. Tree species classification results (D1, sampling rate 100%), where Num refers to the number of these species in the validation set. The sum of correct classification trees (SCI) is the number of correctly classified trees in a validation set of 133.
Table 6. Tree species classification results (D1, sampling rate 100%), where Num refers to the number of these species in the validation set. The sum of correct classification trees (SCI) is the number of correctly classified trees in a validation set of 133.
NameM.F n = 16C.C n = 19C.S
n = 8
A.B n = 21S.B
n = 23
Gl.
n = 12
M.G
n = 8
G.B n = 26UA (%)CE
(%)
Num.
M.F12010000185.714.3
C.C11910202082.617.4
C.S00700000100.00
A.B000200000100.00
S.B302021000100.00
Gl.0000090190.010.0
M.G000000601000
G.B10000212385.114.9
PA (%)85.795.087.595.291.375.075.088.5OA: 86.7%
OE (%)14.35.012.54.88.725.025.011.5SCI: 117
Table 7. Tree species classification results (D2, sampling rate 80%), where Num refers to the number of these species in the validation set. The sum of correct classification trees (SCI) is the number of correctly classified trees in a validation set of 133.
Table 7. Tree species classification results (D2, sampling rate 80%), where Num refers to the number of these species in the validation set. The sum of correct classification trees (SCI) is the number of correctly classified trees in a validation set of 133.
NameM.F n = 16C.C n = 19C.S
n = 8
A.B n = 21S.B
n = 23
Gl.
n = 12
M.G
n = 8
G.B n = 26UA (%)CE
(%)
Num.
M.F10110010076.923.1
C.C11500000093.36.7
C.S0140000080.020.0
A.B00019000195.05.0
S.B10001710088.012.0
Gl.1000070363.656.4
M.G0100006085.714.3
G.B00000101995.05.0
PA (%)62.578.950.090.595.758.375.084.6OA: 87.4%
OE (%)37.521.150.09.54.341.725.015.4SCI: 97
Table 8. Tree species classification results (D3, sampling rate 50%), where Num refers to the number of these species in the validation set. The sum of correct classification trees (SCI) is the number of correctly classified trees in a validation set of 133.
Table 8. Tree species classification results (D3, sampling rate 50%), where Num refers to the number of these species in the validation set. The sum of correct classification trees (SCI) is the number of correctly classified trees in a validation set of 133.
NameM.F n = 16C.C n = 19C.S
n = 8
A.B n = 21S.B
n = 23
Gl.
n = 12
M.G
n = 8
G.B n = 26UA (%)CE
(%)
Num.
M.F7000000187.512.5
C.C31210000266.733.3
C.S1030000075.025.0
A.B000200001100.00
S.B1000800088.012.0
Gl.1000030075.025.0
M.G0000004180.020.0
G.B01100101785.015.0
PA (%)62.563.237.595.234.725.050.065.4OA: 84.1%
OE (%)37.536.862.54.865.375.050.034.6SCI: 74
Table 9. Classification results of tree species (D4, sample rate 30%), where Num refers to the number of these species in the validation set. The sum of correct classification trees (SCI) is the number of correctly classified trees in a validation set of 133.
Table 9. Classification results of tree species (D4, sample rate 30%), where Num refers to the number of these species in the validation set. The sum of correct classification trees (SCI) is the number of correctly classified trees in a validation set of 133.
NameM.F n = 16C.C n = 19C.S
n = 8
A.B n = 21S.B
n = 23
Gl.
n = 12
M.G
n = 8
G.B n = 26UA (%)CE
(%)
Num.
M.F3100000075.025
C.C0700000277.822.1
C.S0240000066.733.3
A.B000190001100.00
S.B00201100084.615.4
Gl.1000010050.050.0
M.G0100004080.020.0
G.B10000001694.15.9
PA (%)18.863.237.590.447.825.08.361.5OA: 85.5%
OE (%)81.236.862.59.652.275.091.738.5SCI: 65
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gong, Y.; Li, X.; Du, H.; Zhou, G.; Mao, F.; Zhou, L.; Zhang, B.; Xuan, J.; Zhu, D. Tree Species Classifications of Urban Forests Using UAV-LiDAR Intensity Frequency Data. Remote Sens. 2023, 15, 110. https://0-doi-org.brum.beds.ac.uk/10.3390/rs15010110

AMA Style

Gong Y, Li X, Du H, Zhou G, Mao F, Zhou L, Zhang B, Xuan J, Zhu D. Tree Species Classifications of Urban Forests Using UAV-LiDAR Intensity Frequency Data. Remote Sensing. 2023; 15(1):110. https://0-doi-org.brum.beds.ac.uk/10.3390/rs15010110

Chicago/Turabian Style

Gong, Yulin, Xuejian Li, Huaqiang Du, Guomo Zhou, Fangjie Mao, Lv Zhou, Bo Zhang, Jie Xuan, and Dien Zhu. 2023. "Tree Species Classifications of Urban Forests Using UAV-LiDAR Intensity Frequency Data" Remote Sensing 15, no. 1: 110. https://0-doi-org.brum.beds.ac.uk/10.3390/rs15010110

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop