Detection of New Zealand Kauri Trees with AISA Aerial Hyperspectral Data for Use in Multispectral Monitoring

The endemic New Zealand kauri trees (Agathis australis) are of major importance for the forests in the northern part of New Zealand. The mapping of kauri locations is required for the monitoring of the deadly kauri dieback disease (Phytophthora agathidicida (PTA)). In this study, we developed a method to identify kauri trees by optical remote sensing that can be applied in an area-wide campaign. Dead and dying trees were separated in one class and the remaining trees with no to medium stress symptoms were defined in the two classes “kauri” and “other”. The reference dataset covers a representative selection of 3165 precisely located crowns of kauri and 21 other canopy species in the Waitakere Ranges west of Auckland. The analysis is based on an airborne hyperspectral AISA Fenix image (437–2337 nm, 1 m2 pixel resolution). The kauri spectra show characteristically steep reflectance and absorption features in the near-infrared (NIR) region with a distinct long descent at 1215 nm, which can be parameterised with a modified Normalised Water Index (mNDWI-Hyp). With a Jeffries–Matusita separability over 1.9, the kauri spectra can be well separated from 21 other canopy vegetation spectra. The Random Forest classifier performed slightly better than Support Vector Machine. A combination of the mNDWI-Hyp index with four additional spectral indices with three red to NIR bands resulted in an overall pixel-based accuracy (OA) of 91.7% for crowns larger 3 m diameter. While the user’s and producer’s accuracies for the class “kauri” with 94.6% and 94.8% are suitable for management purposes, the separation of “dead/dying trees” from “other” canopy vegetation poses the main challenge. The OA can be improved to 93.8% by combining “kauri” and “dead/dying” trees in one class, separate classifications for low and high forest stands and a binning to 10 nm bandwidths. Additional wavelengths and their respective indices only improved the OA up to 0.6%. The method developed in this study allows an accurate location of kauri trees for an area-wide mapping with a five-band multispectral sensor in a representative selection of forest ecosystems.


Introduction
New Zealand kauri (Agathis australis (D.Don) Lindl. ex Loudon) are an important component of New Zealand's northern indigenous forests. The overall distribution of kauri is well known [1,2], but there is an urgent need to locate kauri crowns in more detail for monitoring the deadly kauri Table 1. Spectral ranges with wavelengths used in this study (adapted from [11]).
Airborne hyperspectral remote sensing has proven useful in the analysis and identification of individual tree crowns in boreal and temperate [12][13][14][15] and subtropical to tropical forests [16][17][18][19][20]. The NIR bands from 700 to 1327 nm are important for species classification in tropical forests [16][17][18], which are perhaps more similar to kauri forests in terms of structural complexity [7]. The reduction of dimensionality and correlation in hyperspectral datasets can significantly improve the accuracy [21][22][23], as well as the extraction of the sunlit part of the crown [24,25]. An object-based classification can increase the accuracy by balancing the within crown variation, noise and illumination effects, and it allows to integrate additional structural and spatial crown statistics [26][27][28]. However, the whole processing chain for individual tree crown identification is complex, processing-intensive and error-prone, which can compromise the advantages of an object-based approach [12,29]. The importance of additional LiDAR data describing height and structural crown characteristics has been confirmed in many studies [13,21,30,31]. However, texture measures can also be integrated based on optical data alone [32].
Airborne multispectral sensors are, for the time being, better suited to cover large areas than hyperspectral sensors, with a wider field of view, a larger signal to noise ratio and a more robust technical setup. They are often operated in combination with LiDAR data for tree species classification [33][34][35]. Fassnacht et al. [34] recommended linking the analytical findings in the hyperspectral space with the operational advantages of multispectral sensors.
Most of the kauri in the study area grow in a more diverse second-growth forest [36]. The young growth form of this evergreen conifer is a conical shape with dense foliage (Figure 1). They often

100
The overall aim of this study was to develop a cost-efficient method to identify the location of 101 kauri trees in New Zealand's kauri forests based on optical remote sensing. The method should be 102 applicable for wall-to-wall large area monitoring with multispectral sensors. Dead and dying trees 103 were mapped in a separate class since it was not possible to define spectrally if these are kauri.

104
Moreover, the management needs to document the location of dead trees before they are overgrown 105 or fallen. The resulting "kauri mask" can then be used for further applications such as a detailed 106 analysis of stress symptoms.

107
The main objectives of this study were: 108 • Objective 1: Identify and compare the spectra of kauri and associated canopy tree species 109 with no to medium stress symptoms and analyse their spectral characteristics and 110 separability.

111
• Objective 2: Identify and describe the best spectral indices for the separation of the three 112 target classes "kauri", "dead/dying trees" and "other" canopy vegetation (see class 113 description below).

114
• Objective 3: Define an efficient non-parametric classification method to differentiate the 115 three target classes that is applicable for large area monitoring with multispectral sensors.

116
We chose a pixel-based approach as it did not require a prior crown segmentation.  (Photos: [39]).
The natural distribution of the endemic New Zealand conifer, kauri, extends over the warm temperate lowland forests of the upper North Island (Figure 2), although its abundance has been severely impacted by logging associated with European settlement [2]. In the remaining patches of mature kauri forest, the upper canopy is dominated by large dome shape kauri with an open crown structure and scattered foliage (Figure 1).

Objectives and Approach
The overall aim of this study was to develop a cost-efficient method to identify the location of kauri trees in New Zealand's kauri forests based on optical remote sensing. The method should be applicable for wall-to-wall large area monitoring with multispectral sensors. Dead and dying trees were mapped in a separate class since it was not possible to define spectrally if these are kauri. Moreover, the management needs to document the location of dead trees before they are overgrown or fallen. The resulting "kauri mask" can then be used for further applications such as a detailed analysis of stress symptoms.
The main objectives of this study were: • Objective 1: Identify and compare the spectra of kauri and associated canopy tree species with no to medium stress symptoms and analyse their spectral characteristics and separability. • Objective 2: Identify and describe the best spectral indices for the separation of the three target classes "kauri", "dead/dying trees" and "other" canopy vegetation (see class description below). • Objective 3: Define an efficient non-parametric classification method to differentiate the three target classes that is applicable for large area monitoring with multispectral sensors.
We chose a pixel-based approach as it did not require a prior crown segmentation.

Study Area
Three sites in the Waitakere Ranges Heritage Area, northwest of central Auckland (Figure 2), cover a representative range of kauri stands in all sizes and stages of stress [36]. The Cascade area (10.3 km 2 ) contains patches of old established kauri stands, the Maungaroa area (5.4 km 2 ) includes mainly second-growth kauri forests, and a diverse selection of mature crowns can be found in the Kauri Grove Valley (1.1 km 2 ). A rough terrain characterises the ranges with elevations from sea level to a maximum of 336 m in the study sites and 474 m at the highest peak [40]. The climate is warm-temperate and influenced by the proximity of the sea [41].

Data and Data Preparation
LiDAR data (RIEGL LMS-Q1560 sensor, average 35 returns/m 2 with circa 0.5 ground returns/m 2 ) and RGB aerial images (15 cm) were flown for the three study sites in one acquisition on 30 January 2016. Pit-free terrain model (DTM), surface model (DSM) and crown-height model (CHM) were generated with LAStools [43]. The aerial image was orthorectified in two versions, on the DTM and the DSM. An additional 7.5 cm RGB aerial image was acquired in summer 2017 [44].
The airborne hyperspectral image was acquired on 15 March 2017 with an AISA Fenix hyperspectral sensor at 1 m pixel resolution and was delivered in 23 stripes in radiance units. The sensor features 448 spectral bands with an average bandwidth of 3.6 nm in the VNIR1 region and 10 nm in the NIR2/SWIR-region. The flight conditions were cloud-free but windy, with a high amount of moisture in the forest after recent rain. Reflectance measurements with an ASD field spectrometer were taken as a reference during the flight on homogenous flat areas (grass, gravel, tarmac) and black and white sheets of 5 m × 5 m.
The atmospheric correction was performed using ATCOR 4 [14] with a variable water vapour correction on the 1130 nm spectral region and a "maritime" atmosphere setting for the aerosol parameters. The spectral distortions of the push broom sensor were addressed by developing a sensor model with an adapted shift in the bandwidths. The parameters for the shift were empirically derived for each sensor part (VNIR1 and SWIR) from atmospheric gas absorption features on a homogenous part of the image. The O 2 absorption bands at 760 and 820 nm could be sufficiently corrected by applying the sensor-shift in ATCOR. Remaining spikes and dips in the 940 and 1130 nm regions were removed by applying a non-linear interpolation. The ASD reflectance field measures were used as a reference to evaluate the parameters for the atmospheric correction, not for the analysis itself.
The original image showed some distinct non-periodic, single black and white "bad lines" in the columns of wavelengths at the beginning and end of the spectrum and close to the shift between the VNIR and SWIR sensors. These lines were identified by their mean value compared to the mean values of the direct neighbouring lines according to a local approach described in [45]. For the de-striping, the pixel values in these lines were replaced with the average of the neighbouring pixels.
The geographic distortions were corrected in a two-step approach: First, the basic corrections for the Global Navigation Satellite System (GNSS) position, altitude, roll, pitch, heading and offset between the inertial measurement unit and lens were applied in PARGE [46]. In a second step, the remaining distortions were corrected with an individual polynomial orthorectification per stripe in ERDAS Imagine based on over 2300 ground control points.
Ninety-six of the 448 bands that were most affected by noise and stripes were removed, leaving 352 useable bands for the analysis. The noisy bands were located in the beginning and end of the spectrum and in the absorption bands of water.
The 23 corrected stripes were stitched together with "mosaic data seamlines" in ArcGIS to three smaller mosaics covering the three study areas before they were combined into one large mosaic covering ca. 9 km 2 .
During fieldwork in the 2015/2016 and 2016/2017 summer months, the reference crowns in denser stands were located with a mapping grade GNSS (Trimble-GeoXH-3.5G) with distance and bearing in circular sampling plots of 20 and 30 m diameter. In open stands, crowns were edited directly on aerial images and a CHM layer on a field tablet. Table A1 gives an overview of the reference data with scientific names and the priority of neighbouring canopy species, according to their resemblance to and association with kauri. A threshold of at least 40% dead branches visible in the crown area in the 2017 aerial image was defined to identify the class "dead/dying trees". The sunlit parts were identified with a threshold on the average of the RGB-NIR bands [47]. The challenge was to define a brightness threshold that removes the core shadow areas without useful spectral information and keep the partly shaded inner-crown pixels that still contribute to the species identification. The threshold on the RGB average was defined by comparing the resulting areas with manually identified sunlit parts of the crown. A brightness threshold was also calculated on the NIR band, to match reduced band selections. Edge effects were reduced by removing an internal buffer of 10% crown diameter. The final reference set includes 3165 crowns with a total of 95,194 sunlit pixels in 1 m 2 resolution (Table A1 and Figure 3).
The crown size classes used in this study refer to the mean crown diameter. It was defined as the average of the maximum and the minimum diameter based on the "minimum bounding geometry -rectangle by width" in ArcGIS. The thresholds for the size classes were empirically defined from the field measurements to mark the transition from small kauri crowns (>3 m to 4.8 m diameter) to the more open medium crown sizes (>4.8 m to <12.2 m diameter) and the large dome shape crowns (>12.2 m diameter) (Figure 1). In addition, the minimum object size for a 1 m pixel resolution was marked in a class of very small crowns of <3 m diameter. The information about the crown size was used to interpret results of the pixel-based classification, not as an attribute in the classification.
As a preparation for a separated analysis for different forest types, two forest stand categories "high" and "low" were segmented on the CHM in eCognition (scale 15 m, shape 0.3 and compactness 0.9 [48]) and defined by a mean height threshold of 21 m.
The crowns and thereby the reference pixels were sorted in three target classes for the analysis: • "dead/dying trees" with a minimum of 40% visible dead branches in the aerial image; • "kauri" that were not classified as "dead/dying"; and • "other" canopy vegetation that was not classified as "dead/dying".
The crowns in the classes "kauri" and "other" showed no to medium stress symptoms with an intact crown architecture.    combinations was included in the analysis. This objective has two tasks: 219 1. separate "dead/dying trees" from less symptomatic "kauri" and "other" canopy vegetation; and 220 2. distinguish "kauri" from "other" canopy vegetation.

Extraction and Analysis of Spectra and Spectral Separabilities
Outliers that were caused by mixed pixels, single dead branches or patches of deviant plant material could be visually identified in Envi's n-D Visualizer with the bands 1, 3 and 5 of a Minimum Noise Transformation (MNF) on all 352 bands [49]. These outlier pixels were removed for each class for test purposes. The mean signatures of kauri and associated tree species, the standard deviation and Jeffries-Matusita separability were calculated both with and without the removal of outlier pixels.
A Random Forest classification of kauri and 10 neighbouring tree species was calculated with a 10-fold cross-validation in 10 random repetitions. Only crowns larger than 5 m diameter with no or slight stress symptoms were included, to reduce the confusion with mixed pixels and declining foliage. A randomly spread subsample was extracted from the more frequent classes to match the species distribution in the study areas. The results were presented in a confusion matrix.

Band and Indices Selection
The aim of the selection process was to identify a set of 4-8 wavelengths and derived indices to distinguish the three target classes. Multispectral sensors usually feature up to six bands, but since an eight-band multispectral sensor was available, a maximum of eight bands for the index combinations was included in the analysis. This objective has two tasks: separate "dead/dying trees" from less symptomatic "kauri" and "other" canopy vegetation; and 2.
Initially, 52 indices were calculated on the 352 selected hyperspectral bands. In noisy areas of the spectrum, the values of three neighbouring bands were averaged. Indices with a high correlation (>0.98 or <-0.98) were removed by reducing the number of bands and keeping the best performing indices. For the attribute selection, several ranker methods (Correlation, GainRatio, InfoGain, Symmetrical Uncertainty and Principal Component) were combined in WEKA [50] by applying a weight according to the ranking results. The final combinations with 4-8 bands were identified with a Wrapper Subset Evaluator and the attribute importance for a Random Forest classification. The same selection process was repeated with indices in only the visible to NIR1 spectral range (VNIR1), up to 970 nm.

Selection and Parametrisation of the Classifier
Random Forest (RF) and Support Vector Machine (SVM) have been used successfully for tree classification in several studies [51][52][53][54]. As non-parametric classifiers, they do not require a normal distribution of the reference data and are well suited to handle a large number of attributes and high variability in the classes [55,56]. The SVM separates the classes by constructing a hyperplane based on support vectors at the outer class edges [57]. The parameters (cost: 1000, gamma: 0.1) were defined in WEKA with the GridSearch package [58]. The Sequential Minimal Optimization function in WEKA for the SVM analysis could handle the three target classes by using pairwise classification.
The RF classifier combines a large number of decision trees based on bootstrap samples with an ensemble learning algorithm [51]. A random selection of a given number of features is used to split each node in the RF implementation in WEKA. The final model is based on the number of similar outcomes ("votes") from all decision trees [59]. The parameters were systematically tested, and the highest accuracies could be achieved with 500 trees, two attributes per node and a maximum tree depth of 40. The performances of both classifiers with the defined parameters were tested in a five-fold random split of all sunlit pixels with 20 repetitions. As expected, the use of alternative classifiers (Maximum Likelihood, J48 decision tree) yielded inferior results in comparison to RF and SVM.

Tests to Further Improve the Accuracy
The default classification was calculated on the defined band selections and parameters. Several tests were conducted to improve the accuracies: • resampling of the original bandwidths to 10 nm, 20 nm and 30 nm; • addition of three selected texture values on the 800 nm band (data range (7 kernel (k)), variance (7 k) and second moment (3 k)), following the procedure for the indices' selection; • addition of a LiDAR CHM as a layer for the classification; • separate classifications for low and high stands; and • removal or reclassification of outlier pixels in the training set.
The final accuracies were calculated pixel-based with test pixels in all crowns. Producer's and user's accuracies were determined for the three target classes as the mean values from all repetitions in the RF classifications.

Results Objective 1: Kauri Spectrum
Compared to the mean spectra of other canopy vegetation, the mean spectra of kauri pixels ( Figure 4) show a slightly lower reflectance in the green part of the spectra and lower signals in all spectral regions. The most distinctive feature in the kauri spectrum is a steep ascend from 1000 nm to 1070 nm with a long descent to the absorption feature at 1215 nm. The bands of the NIR2 range are the most important for kauri identification, followed by the NIR1 and SWIR1 (Figures 4 and 5). The spectra of very small kauri crowns (<3 m DM) differ slightly from larger kauri (>4.8 m DM) with a Transformed Divergence value of 1.95 [60].

277
With a Jeffries-Matusita value of over 1.9 [61], the pixel-based spectra of 21 other species can all 278 well be separated from the pixel-based kauri spectra (Table B1). The separability increases after the 279 removal of outlier pixels. The main species that are incorrectly classified as kauri are rimu, tanekaha, 280 rewarewa, tōtara, miro and kawaka (Table B1). These species show similar spectral features to the 281 kauri spectrum with the long descent in the NIR2 range and lower SWIR values ( Figure 6). The 282 spectra of species with a high spectral separability from kauri, such as flax, kanuka, tree fern and

277
With a Jeffries-Matusita value of over 1.9 [61], the pixel-based spectra of 21 other species can all 278 well be separated from the pixel-based kauri spectra (Table B1). The separability increases after the 279 removal of outlier pixels. The main species that are incorrectly classified as kauri are rimu, tanekaha, 280 rewarewa, tōtara, miro and kawaka (Table B1)  With a Jeffries-Matusita value of over 1.9 [61], the pixel-based spectra of 21 other species can all well be separated from the pixel-based kauri spectra (Table A2). The separability increases after the removal of outlier pixels. The main species that are incorrectly classified as kauri are rimu, tanekaha, rewarewa, tōtara, miro and kawaka (Table A2). These species show similar spectral features to the kauri spectrum with the long descent in the NIR2 range and lower SWIR values ( Figure 6). The spectra of species with a high spectral separability from kauri, such as flax, kanuka, tree fern and pohutukawa have higher reflectance features in the VIS, NIR and SWIR range and a lower descent to the 1215 nm water vapour window (Figure 7). Remote Sens. 2019, 11, x FOR PEER REVIEW 9 of 29 286 Figure 6. Mean spectra of kauri (thick black line) and six selected other canopy species (grey) that got 287 most easily confused with kauri. The number of pixels (pix) used to generate the mean spectra is given 288 in parentheses. The spectra of these species show the lowest separability from the kauri spectrum in 289 this study (see Table B1). separabilities from the kauri spectrum in this study (see Table B1). The number of pixels (pix) used to 293 generate the mean spectra is given in parentheses.

294
The overall high separability of kauri with neighbouring species could also be confirmed in a 295 classification of kauri and 10 other tree species on the full spectral range of the AISA image (Table 2).

296
Only non-symptomatic crowns larger 5 m diameter were chosen to avoid confusion caused by mixed 297 pixels and declining foliage. The overall accuracy of 94.8% and user's accuracies from 98.1% for rata 298 to 98.7% for kauri, confirm the high spectral separability of kauri and also between the selected 10 299 tree species. Most species show high producer's accuracies of over 93% with 99.1% for kauri.
302 Figure 6. Mean spectra of kauri (thick black line) and six selected other canopy species (grey) that got most easily confused with kauri. The number of pixels (pix) used to generate the mean spectra is given in parentheses. The spectra of these species show the lowest separability from the kauri spectrum in this study (see Table A2).
Remote Sens. 2019, 11, x FOR PEER REVIEW 9 of 29 286 Figure 6. Mean spectra of kauri (thick black line) and six selected other canopy species (grey) that got 287 most easily confused with kauri. The number of pixels (pix) used to generate the mean spectra is given 288 in parentheses. The spectra of these species show the lowest separability from the kauri spectrum in 289 this study (see Table B1).
290 291 Figure 7. Mean spectra of kauri (black) and five other canopy species (grey) that have the highest 292 separabilities from the kauri spectrum in this study (see Table B1). The number of pixels (pix) used to 293 generate the mean spectra is given in parentheses.

294
The overall high separability of kauri with neighbouring species could also be confirmed in a 295 classification of kauri and 10 other tree species on the full spectral range of the AISA image (Table 2).

296
Only non-symptomatic crowns larger 5 m diameter were chosen to avoid confusion caused by mixed 297 pixels and declining foliage. The overall accuracy of 94.8% and user's accuracies from 98.1% for rata 298 to 98.7% for kauri, confirm the high spectral separability of kauri and also between the selected 10 299 tree species. Most species show high producer's accuracies of over 93% with 99.1% for kauri.

300
However, tōtara, rewarewa, tanekaha and miro have the lowest producer's accuracies: from 58% for 301 rewarewa to 77% for miro. Figure 7. Mean spectra of kauri (black) and five other canopy species (grey) that have the highest separabilities from the kauri spectrum in this study (see Table A2). The number of pixels (pix) used to generate the mean spectra is given in parentheses.

302
The overall high separability of kauri with neighbouring species could also be confirmed in a classification of kauri and 10 other tree species on the full spectral range of the AISA image (Table 2). Only non-symptomatic crowns larger 5 m diameter were chosen to avoid confusion caused by mixed pixels and declining foliage. The overall accuracy of 94.8% and user's accuracies from 98.1% for rata to 98.7% for kauri, confirm the high spectral separability of kauri and also between the selected 10 tree species. Most species show high producer's accuracies of over 93% with 99.1% for kauri. However, tōtara, rewarewa, tanekaha and miro have the lowest producer's accuracies: from 58% for rewarewa to 77% for miro. Table 2. Confusion matrix and user's and producer's accuracies for a RF classification of kauri and ten neighbouring tree species on the full hyperspectral range of the AISA image (first 25 bands of a 35 band MNF transformation) evaluated with a 10-fold cross-validation for the seed value 1. Only sunlit pixels of trees with a minimum diameter of 5 m were chosen to avoid shadows and to reduce the effects of mixed pixels. The selected crowns were either non-symptomatic or showed only mild symptoms of stress. The category "dead/dying" was difficult to define from the classes "kauri" and "other" with user's accuracy of 80.3% and a producer's accuracy of only 52.1% in the final setup. In a test with aggregated percentages of the classes per reference crown, the producer's accuracy for the class "dead/dying" could be improved to 75.5% for a minimum threshold of 15% crown area defined as "dead/dying". The main characteristic features of the spectra of dead/dying trees are a lower chlorophyll absorption in the red region (around 670 nm), a lower reflectance of green leaf scattering in the NIR1 region (around 800 nm), a blue shift of the red edge point and overall high values in the SWIR region ( Figure 5). Tests with separate classes for incorrectly classified pixels as well as the inclusion of shadow pixels gave no improvement for the "dead/dying" class.

Results Objective 2: Indices Selection
A preselection of 13 best performing indices over the whole spectrum is described in Error! Reference source not found., and their position in relation to the mean spectra of the target classes is illustrated in Figure 8. Figure 9 presents the performance of each index to identify the class "dead/dying" and to distinguish "kauri" from "other" canopy vegetation, with the best resulting combinations shown in Figure 10. A paired t-test for the resulting accuracies with a p-value of 0.05 confirmed that these results and thereby the ranking of index combinations are significant.
For a four-band multispectral sensor, the highest performance of 90.1% OA ( Figure 10) could be achieved with four indices, based on bands in the VIS (670 nm), NIR1 (800 nm) and NIR2 region (1074 and 1209 nm). The combination of three indices on the red and NIR1 bands helped to identify the class "dead/dying" (Figure 9). The NIR 2 spectral range proved to be the most important for the identification of kauri followed by the NIR1, SWIR1, VIS and SWIR2 spectral ranges (Figures 4 and 11).

321
A preselection of 13 best performing indices over the whole spectrum is described in Error!

322
Reference source not found., and their position in relation to the mean spectra of the target classes is 323 illustrated in Figure 8. Figure 9 presents the performance of each index to identify the class 324 "dead/dying" and to distinguish "kauri" from "other" canopy vegetation, with the best resulting 325 combinations shown in Figure 10. A paired t-test for the resulting accuracies with a p-value of 0.05 326 confirmed that these results and thereby the ranking of index combinations are significant.

336
The best distinction between kauri and other canopy vegetation could be achieved with a 337 normalised index (mNDWI-Hyp, Figure 12) that captures the distinctive long descent in the NIR2 338 spectrum. It was first described as an alternative to a Normalised Water Index (NDWI) adapted to

373
(light grey) and to distinguish between "kauri" and "other vegetation" (dark grey) with an RF 374 classification (five-fold random split, 20 repetitions). Please note that the x-axis starts at 55%.

378
Please note that the y-axis starts at 89%.   the NDNI index (see Table C1 for descriptions of these indices).

390
In general, the NIR2 Indices are more important to distinguish kauri than indices in the visible 391 to NIR1 (VNIR1) range. The best performing VNIR1 index combination for an eight-band sensor 392 includes bands in the 550-970 nm spectral range (Table 3). This combination resulted in 84.6% OA to The best distinction between kauri and other canopy vegetation could be achieved with a normalised index (mNDWI-Hyp, Figure 12) that captures the distinctive long descent in the NIR2 spectrum. It was first described as an alternative to a Normalised Water Index (NDWI) adapted to Hyperion data [59] and was further modified in this study by using the natural logarithm values to address outliers (after [60]).

378
Please note that the y-axis starts at 89%.   Table C1 for descriptions of these indices).

390
In general, the NIR2 Indices are more important to distinguish kauri than indices in the visible 391 to NIR1 (VNIR1) range. The best performing VNIR1 index combination for an eight-band sensor 392 includes bands in the 550-970 nm spectral range (Table 3). This combination resulted in 84.6% OA to 393 Figure 12. Histograms for selected indices on sunlit pixels for all crown diameters, with the class "kauri" marked in light blue, the class "dead/dying" in red and the class "other" in dark blue. (a) The histogram for the mNDWI-Hyp index, which performed best to separate the class kauri from other vegetation by capturing distinctive features in the NIR2 region, is shown. For the separation of the class "dead/dying", indices in the RED/NIR1 region are better suited, such as (b) the SR800 and (c) the NDNI index (see Table A3 for descriptions of these indices).
Other indices that are useful to identify kauri like the Moisture Stress Index (MSI), NDWI and Water Band Index (WBI) (Figure 9) also include bands in the NIR1 and NIR2 spectral range. However, in combination with the best performing mNDWI-Hyp index, they did not increase the overall accuracy.
For a five-band sensor, an additional Simple Ratio Index with an extra Red Edge band at 708 nm (SR708) increased the OA to 90.8% for all classes on all crown sizes. The combination of the SR708 with the other three indices on the RED to NIR1 bands performed best to distinguish the class "dead/dying" Figure 9. This five-band combination was considered the best trade-off between the number of bands and the resulting accuracy. It was therefore used as the default combination for the development of the finale classification method. The inclusion of further bands and respective indices only resulted in slight improvements in the accuracy.
An additional band at 970 nm for a six-band sensor allows including a further Normalised Difference Index (ND970) and resulted in an OA of 90.9% (Figure 10). This index was developed in this study to describe the characteristically steep ascend in the kauri spectra from the first NIR water vapour window at 970 nm to the reflectance feature on 1074 nm ( Figure 5, Table A3).
With seven multispectral bands, the best results of 91.3% OA could be achieved by adding the Normalised Difference Nitrogen Index (NDNI) with two bands in the SWIR1 region ( Figure 9, Table A3). It describes the leaf nitrogen concentration in the 1510 nm band in relation to the canopy foliar mass measured at 1680 nm, which again depends on the absorption by leaf and canopy water [62]. The kauri spectrum shows a lower magnitude in the slope between the 1510 nm band versus the reflectance feature at 1680 nm compared to the mean spectrum of the two other target classes (see Figure 5).
As an alternative for a seven-band sensor, the addition of a Photochemical Reflectance Index (PRI) [63], with two bands in the green region, results in an OA of 91.2% (StD 0.19). This index describes the photosynthetic light use efficiency by carbon dioxide uptake. It captures the slightly lower green reflectance feature in the kauri spectrum. A test confirmed its usability on a resampled 10 nm bandwidth.
With eight spectral bands available, the highest OA of 91.3% (StD 0.2) could be achieved by adding both the 970 nm band for the ND970 index and the two SWIR1 bands for the NDNI index to the five bands of the default setup.
In general, the NIR2 Indices are more important to distinguish kauri than indices in the visible to NIR1 (VNIR1) range. The best performing VNIR1 index combination for an eight-band sensor includes bands in the 550-970 nm spectral range (Table 3). This combination resulted in 84.6% OA to distinguish the three target classes ( Figure 13). If only six bands are available, three indices on red to NIR1 bands (675-970 nm) classified the three target classes with an OA of 78.4%. (1) Selection with three indices; (2) The original wavelengths of the index was slightly modified to reduce the number of bands.  (1) Selection with three indices.

398
(2) The original wavelengths of the index was slightly modified to reduce the number of bands.

Results Objective 3: Method Development
The final accuracies are based on an image with five wavelengths (10 nm bandwidth) and five derived indices including the NIR2 bands, according to the recommended index selection for the whole spectrum in the previous sections. It enabled the distinction of "kauri and dead/dying trees" from "other canopy vegetation" (two classes) with a pixel-based overall accuracy (OA) of 93.4%. The three classes with "dead/dying crowns" as a separated category could be identified with 91.3% OA (Table 4, Test E). The separation of the class "dead/dying" from the class "other" poses the main challenge, while the pixel-based user's and producer's accuracies for the class "kauri" are close to 95% (Table 4). These results are based on a RF classification. The RF classifier performed with 90.9% OA for the default setup slightly better than the SVM classifier (89.5% OA), at half of the processing time and it was easier to optimise. The resulting maps for the final setup that was applied to independent test crowns for the three study areas are shown in Figure 14. Crowns that were not chosen as test crowns in the 10 repetitions are marked as "unclassified".
The accuracies for index combinations that include only bands in the visible to NIR1 range are significantly lower with 84.6% OA for three classes on eight bands and 78% on six bands. Combining the classes "kauri" and "dead/dying" improved the OA to 86.9% for the eight-band selection (see Figure 13). Further improvements of about 2% could be achieved by adding a CHM layer.
The full spectral range of 25 MNF bands resulted in overall pixel-based accuracy of 93.9% for three classes and 96.2% for two classes. Attempts to remove mixed pixels and noise by excluding an MNF forward and backward transformation did not improve the overall accuracy.
A binning to 10 nm helped to remove noise and redundancies (Table 4, Tests B1 and B2), while 20 and 30 nm resampling proved to be too coarse to capture the small spectral windows of the selected indices.
A separated classification for low and high stands ( Table 4, Test C) improved the accuracy by 1.5%. Adding a CHM layer achieved a similar improvement, but it was not used for the final setups, because the LiDAR data do not match the hyperspectral image sufficiently for a direct pixel-based combination. Additional texture features based on the 800 nm NIR band gave a slight improvement in the classification of small crowns but lowered the overall accuracy in the larger crowns. In addition, the partial removal of outliers in the training set (Table 4, Test D) resulted in a slightly enhanced OA of 0.7%. This method was not applied for the final accuracies because it is too elaborate for large area applications.
Post-processing by reclassifying kauri pixels with a height lower than 4 m to the class "other" improved wrongly classified lower shrub areas, but it requires a spatially matching CHM. The merging of singular pixels with a majority kernel according to the stand situation improved the pixel-based accuracy and should be considered for further analysis.

Discussion and Recommendations for Further Analysis
The use of a multispectral sensor with at least five bands in the VIS to NIR2 range is recommended for the detection of kauri and dead/dying trees. This study confirms the findings of Asner [68], Clark, Roberts [69] and Ferreira, Zortea [20] about the importance of the NIR spectrum for the identification of tree species in a diverse forest environment.
Index combinations with bands only in the visible to NIR1 range (up to 970 nm) perform significantly lower than index combinations that include NIR2 and SWIR bands. The overall accuracy was only 84.6% for three target classes on all crown sizes in the visible to NIR1 range, compared to 89.9% accuracy for a similar setup that includes bands in the NIR2 spectrum. If only bands in the VNIR1 spectral range are available, a combination with LiDAR attributes is recommended, ideally in an object-based approach according to the authors of [26][27][28].
The characteristic high reflectance in the kauri spectrum at 1070 nm indicates a particularly high amount of scattering of radiation at air-cell-water edges in the complex structure of the kauri foliage and the thick kauri leaves [11,62,70]. The pronounced water vapour window at 1215 nm is caused by a strong absorption from high leaf, respective crown water content. These results confirm the field observations that kauri crowns are more distinct in structural features than in colour. Since there was a lot of moisture in the forest on the flight day of the AISA sensor, the performance of the selected indices should also be tested under dryer conditions.
Other well-performing indices to identify kauri such as the MSI and NDWI also have bands in the NIR1 and NIR2 range and confirm the importance of structural features and water content for kauri identification. The lower reflectance values of kauri in all spectral regions is most likely caused by the more open crown structure in medium and large kauri compared to neighbouring species.
The main species that are incorrectly classified as kauri tend to have a similar "rough" foliage or needle-like leaves such as rimu, tanekaha, rewarewa, tōtara, miro and kawaka (Table 4, Test B1). Species with similar conical shapes in smaller growth stages such as tanekaha, rimu, kahikatea and rewarewa are easily confused with small kauri. They show low producer's accuracies from 58% for rewarewa to 77% for miro in the individual species classification ( Table 2). While rata has overall high user's and producer's accuracies of 89.0% and 97.8%, it has wrongly classified pixels with all other tree species, including kauri. This confusion is most likely caused by the fact that rata starts its growing cycle as an epiphyte and occurs therefore as part of the foliage on other trees.
The category "dead/dying" is difficult to define because of the graded transition from the two other classes for trees with declining foliage and a higher amount of shadow and mixed reflectance with understory layers. In addition, canopy vegetation with a high amount of carbon fibre such as flax and cabbage trees, wooden seed capsules on kānuka and older dry foliage on rimu were wrongly classified as "dead/dying". In addition, specular reflections on the smooth waxy surface of kahikatea trees and the shiny leaves of tree ferns cause confusion with the class "dead/dying". Higher producer's accuracy in a test with a crown-aggregated setup revealed that misclassification of the class "dead/dying" is partly caused by single pixels on dead branch material in otherwise less symptomatic crowns. While the classification of these pixels is correct, they appear as wrongly classified in the confusion matrix because the reference is crown based.
A separated classification for low and high stands ( Table 4, Test C) improved the OA about 1.5%. This can be explained by a reduced variability in the dataset after separating young trees with dense foliage and lower shrub layers from the mature trees in the higher stands. An alternative to consider different size classes is the direct inclusion of a CHM as an additional layer. For a pixel-based classification, this requires a sub-pixel matching between the optical data and the CHM, which is difficult to achieve in a varied topography with large trees.
The partial removal of outliers in the training set (Table 4, Test D) reduced the effect of mixed pixels, especially for small crowns and resulted in a slightly enhanced OA by 0.7%. However, this analysis is elaborate and should only be considered if it is not possible to include LiDAR data or obtain optical data in a higher spatial resolution, which will reduce the number of mixed pixels.
The 1 m pixel size of the AISA Fenix image put some constraints on the analysis of crowns with a diameter smaller than 3 m, with an overall accuracy of 66.6% in the final setup (Table 4, Test E). The identification of small crowns requires a higher spatial resolution, ideally ≤30 cm, to avoid the effect of mixed pixels.
While some spaceborne hyperspectral sensors cover the recommended bands in the NIR2, their spatial resolution of, e.g. 30 m for the Prisma [71] and the EnMAP mission [72] is too coarse for individual tree crown identification. For larger pixel sizes, also in Landsat and Sentinel satellite images, the detection of stands with younger kauri trees should be further investigated with a spectral unmixing approach for homogenous stand units in combination with LiDAR attributes. The potential of the bright green spring aspect for kauri identification could be analysed in a time series of high-resolution satellite data.
The Random Forest classifier is very efficient to handle classes with a high spectral variability; however, the resulting model is difficult to understand. The clear separation of the "kauri" class in the histogram of the mNDWI index ( Figure 12) indicates that a manual decision tree can be developed, which would be easier to understand and to implement.
While the large reference dataset of kauri in different growth and symptom stages is representative for the Waitakere Ranges, the indices and model for kauri identification should be tested and if necessary readjusted in other kauri forests with a different amount and composition of neighbouring species.

Conclusions
This study is the first to analyse the spectra of kauri and the main neighbouring canopy tree species with an airborne hyperspectral sensor on the full VIS to SWIR spectral range. The main objectives were: (1) to describe the kauri spectra and analyse its separability from other neighbouring tree species; (2) to identify the best spectral indices to separate the class "kauri" from "other" and "dead dying" canopy vegetation; and (3) to define a method for classification of the three target classes that is applicable for large area monitoring with multispectral sensors.
Kauri crowns have characteristic spectra with a steep reflectance feature in the NIR2 spectral region at 1070 nm and a distinct descent to the water vapour windows at 1215 nm and lower reflectance features in the green and SWIR spectral region than other canopy vegetation. The spectral characteristics indicate that kauri crowns are more distinct in their structural than biochemical features. The high separabilities of the kauri spectra from 21 other tree species and canopy vegetation with a Jeffries-Matusita separability larger 1.9 could be confirmed with a high OA of 94.8% for the classification of non-symptomatic crowns larger 5 m diameter of kauri and 10 other tree species.
For the use on a five-band multispectral sensor, five indices (Table A3) in the VIS to NIR2 range performed best to distinguish the three target classes "kauri", "dead/dying trees" and "other canopy vegetation". They are suitable for multispectral area-wide forest mapping.
The Random Forest classifier performed slightly better than Support Vector Machine. The final results with 91.7% OA are based on a separated Random Forest classification of low and high forest stand, a binning to 10 nm bandwidth and the removal of very small crowns (<3 m diameter). The class "kauri" could be discriminated with high user's and producer's accuracies of 94.6% and 94.7% from other canopy vegetation by using the selected five bands from the red spectrum at 670-1215 nm in the NIR2 spectrum. The main challenge was the confusion between the classes "dead/dying" and "other" canopy vegetation. A further improvement to 93.8% OA could be achieved by combining "kauri" and "dead/dying" trees in one class as a "kauri mask" for the further analysis, e.g. of stress symptoms. Additional indices enhance the overall accuracy only slightly, up to 0.6% for an eight-band sensor.
The method for accurate, cost efficient, wall-to-wall mapping of kauri trees presented in this study has important implications for the monitoring of kauri dieback disease and the implementation of measures to control disease over the entire distribution of New Zealand's native kauri forests.

Funding:
The Ministry of Primary Industries funded most of the remote sensing data (agreement No. 17766), while the University of Canterbury, the University of Trier and FrontierSI (former CRCSI) Australia provided scholarships for living costs, fieldwork, equipment and additional LiDAR data. Digital Globe and Blackbridge helped with grants for satellite data. Auckland Council supported the fieldwork and supplied LiDAR data and aerial images and Landcare Research provided field equipment. Rapidlasso and Harris Geospatial helped with grants for software licenses. Henning Buddenbaum was supported within the framework of the EnMAP project (FKZ 50 EE 1530) by the German Aerospace Center (DLR) and the Federal Ministry of Economic Affairs and Energy. The publication was funded by the Open Access Fund of Universität Trier and the German Research Foundation (DFG) within the Open Access Publishing funding programme. Table A2. Spectral separability and confusion of the class "kauri" with "other tree species". The Jeffries-Matusita separability is given both for all sunlit pixels and a pixel set with removed outliers. A value over 1.9 indicates a high spectral separability.  [80] (1) The value for the R800 band was averaged with the values of the two neighbouring bands, to reduce noise.