Introduction

Forests provide timber and non-timber products, habitats for a diverse range of flora and fauna, as well as social benefits such as shelter, food and employment. Managing forests to optimise one or more of these benefits is challenging as they are dynamic and undergo continual change from both natural and human-induced afforestation and deforestation [1]. Effective forest management requires comprehensive forest data for a range of temporal and spatial scales. However, obtaining full descriptive data from ground-based approaches is often unfeasible [2]. Remote sensing, including aerial photography, satellite imagery, and light detection and ranging (LiDAR), has increasingly been used as a means of cost-effectively capturing forest data [2, 3].

Aerial photography is the most widely used remote-sensing technique in forest management [4]. Image characteristics such as shape, pattern and texture are used to delineate forest stand boundaries [5], estimate tree height and crown diameter [6] and identify species [7]. Kovats [8] demonstrated that tree height can be measured using high-resolution stereoscopic aerial photos with photogrammetric techniques to measure the lengths of shadows projected onto the ground. Likewise, Dandois and Ellis [9] accurately measured tree height on aerial photos using a computer vision technique. However, these techniques have not been applied operationally as they are dependent on a number of factors including presence of open flat ground and specified sun elevation and angle. Furthermore, subjective interpretation often results in inconsistent degrees of precision and accuracy [10].

Apart from aerial photography, research has increasingly employed satellite imagery in forest assessment. Satellite multi- and hyperspectral sensors capture the electromagnetic radiation emitted by the sun and reflected by the earth’s surface [11]. Spectral and contextual attributes derived from satellite sensors can be modelled against empirically derived biophysical features of the forest, such as stand basal area, height and crown closure [12, 13], stand density [14], leaf area index (LAI) [1517] and forest volume and biomass [1820]. Development of textural analysis such as the grey-level co-occurrence matrix (GLCM) has also allowed the use of additional information from optical sensors to assess forest stand variables such as height, basal area and stand density [21]. Recently, high-resolution satellite imagery (sub-metre) has been used to measure forest structural variables at a finer scale, which makes possible automated detection of individual trees via crown delineation, and modelling individual tree crown size, height, diameter, volume, age class and species composition [3, 2123].

Advantages of optical sensors include large coverage of forest area, easy access and low cost [24••]; consequently, data from optical sensors are useful for large-scale forest assessment and monitoring [25]. The key limitation of passive optical sensors is that they cannot capture detail below the forest canopy and hence cannot directly measure vegetation structure [26].

LiDAR, an active sensor, has been increasingly used in forest assessment. LiDAR directly measures canopy height, which is a commonly used attribute to describe forest structure [27]. Allometric relationships between tree height and diameter at breast height (DBH) have been derived for various tree species [2830], and these two variables are commonly used to derive volume [31]. LiDAR-measured heights, together with other LiDAR-derived metrics (e.g. penetration metrics), are important for modelling other structural attributes, such as canopy cover [32], stem density [33], basal area [34], biomass [35, 36], volume [37] and LAI [38].

LiDAR is demonstrably superior to passive optical sensors for assessing forest structural variables, especially canopy height [39, 40]. Discrete return airborne LiDAR systems now collect stand-level and regional wall-to-wall forest structure attributes (tree height, stand volume and basal area) in the national inventory programmes of Finland, Norway and Sweden [36, 41]. Despite these examples, the high cost of LiDAR acquisition and lack of processing expertise generally preclude wall-to-wall LiDAR surveys for operational use [4, 42].

In order to overcome the limitations of individual sensors and optimise the advantages of different types of sensors, discrete return LiDAR and optical sensors have been integrated to provide more comprehensive and accurate characterisation of forest structure and dynamics [27]. The intention is to keep the acquisition cost low [43] while using the relationship between LiDAR and optical data to better describe a larger extent of the forest resource [44]. Combined use of optical images and LiDAR in forestry utilises both the spectral information in optical imagery and the 3-D information in LiDAR. Therefore, when combining LiDAR and optical spectral sensors, additional information from both sensors is integrated and more accurate results are expected.

Since optical sensors and discrete return LiDAR have been extensively studied and applied in forestry, a number of reviews on their individual use have been conducted [2, 3, 4548, 49••, 50, 51]. However, no comprehensive review of the fusion of discrete return LiDAR with optical sensors has been conducted despite more than a decade having passed since the first integrated study of discrete return LiDAR and Landsat ETM+ [24••]. We address this gap in the literature by reviewing research into the combined use of discrete return LiDAR and optical sensors across a broad range of applications in forestry, specifically with respect to forest delineation and classification, and estimating forest age, species and forest structural variables. This review is limited to sensor fusion studies where multi-sensors are used simultaneously to assess forest attributes.

Fusion of LiDAR and Optical Sensors

Forest Delineation and Classification

Forest classification and forest boundary delineation are important in both natural and planted forest management in assessing forest types and areas, as understanding forest area and location is fundamental for a broad field of applications and users [52]. Aerial photography is most commonly used to determine forest area through manual interpretation. However, forest stand delineation based on manual interpretation of aerial photos can be highly subjective and time-consuming. Forest cover type classification can be achieved by automated image classification by assigning forest cover types and estimating forest variables based on the spectral, textural and auxiliary information in the image. This produces a more objective delineation and reduces time and associated costs [53].

LiDAR data adds a new source of information for forest classification and delineation through direct estimation of forest canopy size and height. Incorporating LiDAR into optical sensor mapping utilises both spectral and structural information to achieve a more accurate forest classification. Nordkvist et al. [54] integrated low-density discrete return airborne LiDAR-derived height metrics with SPOT 5 HRG spectral information for vegetation classification in Sweden and achieved 16.1 % improvement in classification accuracy compared to using SPOT only. The study also compared maximum likelihood and object-based decision tree classification approaches and found the best accuracy was achieved by decision tree classification. Sasaki et al. [55•] achieved a minor improvement in overall land cover classification accuracy (95 to 97.5 %) by an object-based classification approach integrating inputs from high-resolution spectral images captured by digital camera with LiDAR-derived metrics including height, ratio, pulse and intensity parameters. Additionally, Bork and Su [56] used a maximum likelihood approach to classify eight vegetation classes. Using only airborne LiDAR inputs achieved lower accuracy than using only a digital multispectral image; however, the fusion of LiDAR and multispectral imagery increased classification accuracy by 15–20 %. A more recent study explored the benefits of incorporating a digital elevation model (DEM) and digital surface model (DSM) derived from LiDAR into tropical forest mapping using SPOT 5 HRG imagery. This approach enhanced image segmentation and successfully differentiated six vegetation classes, producing an overall forest classification accuracy of 91 % [57•].

The combination of aerial photography and LiDAR has also commonly been used for forest delineation. Wang et al. [58] developed an approach to automatically delineate forest boundaries using both aerial photography and a LiDAR-derived canopy height model (CHM) for the National Forest Inventory of Switzerland. The approach involved two processes: (1) detecting forests using a moving window over the CHM and delineating forest boundaries by analysing the CHM-derived curvature value, vegetation index and (2) using textural values of segmented image objects. Although there are no statistical improvements calculated, the visual results relative to manual digitisation looked very promising. Haywood and Stone [59] developed an automated approach that transforms aerial photos and LiDAR CHM into vectorised forest stand boundaries (65 % overall accuracy). The study did not explicitly show an improvement using sensor fusion, but it demonstrated the possibility of operationalising the fusion of optical sensor and LiDAR in natural forests stand delineation.

A summary of recent studies using integrated LiDAR and optical sensors is shown in Table 1. All of these studies have either shown an improvement (up to 20 %) in forest classification results compared to using a single sensor or made it possible to discriminate further forest classes that otherwise could not be identified by single sensors. All studies reviewed in this section used discrete return LiDAR and other optical sensors. LiDAR data has been acquired at point densities ranging from 0.54 to 11.3 points/m2, with all but one study reporting point density lower than 2 points/m2. It appears that low-density LiDAR is sufficient to classify forest types and delineate forest boundaries when integrated with optical sensors, as only the interpolated LiDAR surfaces are used as inputs.

Table 1 Recent studies of forest classification and delineation using integrated LiDAR and optical sensors

When a fusion approach is used, the common approach to classify forest types and delineate forest boundaries is image classification based on inputs derived from both LiDAR and optical sensors. Machala and Zejdova [60] listed 26 customised arithmetic features derived from discrete return LiDAR and multispectral sensor data that are useful inputs for classifying forest covers. The inputs are commonly rasterised and applied with automated image classification analysis. There is a clear trend towards the use of object-based image analysis (OBIA) rather than pixel-based classification like the maximum likelihood approach. Pixel-based image classification that ignores spatial association among pixels, tends to be sensitive to spectral variations; hence, it is likely to result in a relatively high level of misclassification [61]. OBIA overcomes the issues by carrying out classification on segmented objects that are similar to real land cover features in size and shape [62]. The approach allows consideration of multiple image elements and scales such as texture, shape and context, as opposed to pixel-based classification that solely relies on the pixel values. Overall, OBIA has been proven to produce more accurate classification results compared to a pixel-based approach with single sensor analysis [63]. This may explain why OBIA is favoured in forest classification and delineation with integrated sensors. The statistics to evaluate the performance of forest type classification include classification accuracy matrices or confusion matrices, which compare the classified classes against reference classification and generate a series of statistics such as overall classification accuracy, producer’s and user’s accuracy and kappa coefficient [64].

Forest Species

Accurate characterisation of species is important in forest management, resource planning and monitoring. Remote sensing technologies have been widely used for forest species classification [22, 65]. The recent development of laser scanning also offers the possibility of automatically identifying individual trees and obtaining height and canopy measurements from them [6668], hence providing the possibility of classifying individual tree species [69].

More accurate species classification is expected from the fusion of LiDAR and other sensors due to the synergy of both structural and spectral data. Holmgren et al. [69] used airborne LiDAR-derived height, canopy and intensity parameters to delineate individual tree crowns and calculate tree crown heights and areas then identified tree species for delineated tree crowns using a maximum likelihood approach by integrating LiDAR data with features from high-resolution digital mapping camera (DMC) digital images. The overall species classification accuracy was 96 %, which was an improvement from using LiDAR (91 %) and digital imagery (88 %) individually. Ke et al. [70] evaluated the combined spectral and textural layers from QuickBird imagery and topography, canopy height and intensity from low-density LiDAR (0.16 points/m2) for forest species classification, using object-based segmentation and machine-learning decision trees. The highest classification accuracy (kappa = 91.6 %) was achieved using both spectral and LiDAR-derived metrics with accuracies that were 20 % greater than those that used individual sensors. Sasaki et al. [55•] classified 16 tree species with high density (11.3 points/m2) airborne discrete return LiDAR and multispectral imagery. Although the object-based decision tree classification produced low overall accuracy (31.5 %), a 17 % improvement was made with inputs from both sensors. Some species such as pine and poplar gained significant improvement (32 and 56 %, respectively) with sensor fusion compared with using digital images only. A more recent study used very high-resolution WorldView-2 images and discrete return airborne LiDAR for object-based species classification in a temperate rainforest in Australia. It utilised the spectral features and GLCM textures from the images and a LiDAR-derived CHM and associated statistics to conduct an object-based decision tree classification on a mixture of natural and plantation forest species. The accuracy for LiDAR-only and image-only species classification was 61 and 70 %, respectively, whereas combined sensors improved species classification to 82 % [71•].

Hyperspectral sensors have also been integrated with LiDAR in forest species classification. Several studies have indicated fusion between hyperspectral and LiDAR data enhances forest species differentiation. Dalponte et al. [72] tested leave-one-out covariance (LOOC), support vector machines (SVM) and k-nearest neighbour (k-NN) classifiers for forest species classification using hyperspectral bands and a LiDAR-derived CHM. The best classification accuracy was found using SVM. In total, 23 tree species classes were successfully identified with the technique, with some classes over 90 % accurate. The overall kappa accuracy using combined sensors was 89.2 %, which was slightly higher than using individual sensors (hyperspectral only: 87.9 %, LiDAR only: 89 %). Jones et al. [73] applied similar datasets and techniques to classify more than ten forest species classes. They found most species classes gained accuracy improvement with fusion of hyperspectral and LiDAR inputs (ranging from 0.3 to 19 %) although the overall accuracy improvement was only minor (1.2 %).

The way that sensor fusion has been performed is that high-resolution optical imagery defines forest stand boundaries and provides spectral separation between different forest species. The addition of LiDAR-derived topographic and height and intensity information further reduces within-class spectral variations caused by topography and enhances variations between species classes as different tree species tend to have different heights [70]. Recent studies integrating LiDAR with other sensors to estimate forest species are shown in Table 2. These studies show that sensor fusion has improved the accuracy of species identification and classification by up to 21 %. It is worth noting that species classification from sensor fusion approaches has yielded high variation in accuracies, ranging from 48.4 to 96 % as a result of large variations in species composition in these studies. Discrete return airborne LiDAR was the most common sensor type and the range of point densities was large, spanning 0.16–50 points/m2. The point density of LiDAR acquisitions for species classification was generally higher than for forest area delineation. As with forest area classification, object-based decision tree classification is predominantly used in species classification with LiDAR and multispectral sensors.

Table 2 Recent studies of forest species identification using integrated LiDAR and optical sensors

Hyperspectral sensors capture finer details in the spectral signature (i.e. narrower spectral bands) than multispectral sensors, which allow more detailed differentiation between similar forest types [72]. Generally, hyperspectral sensors alone have shown promising capability in species classification probably due to high spectral resolution. This explains why only minor improvements in species classification accuracy were observed when LiDAR was added (Table 2). Despite high species classification accuracy using only hyperspectral imagery (relative to multispectral imagery), there is no evidence to suggest that fusion of LiDAR and hyperspectral sensors classifies species more accurately than fusion of LiDAR and multispectral sensors. The processing of hyperspectral data can add complexity as they contain a vast array of spectral bands, so a feature selection process is required to eliminate any redundant bands [72]. Rather than conventional classification approaches, non-parametric classifiers, especially SVM has been used when classifying hyperspectral data. SVM is a linear binary classifier that assigns a given test sample from one of the possible labels [74]. The application of SVM has improved classification accuracy, analysis time and stability with hyperspectral data [73].

Tree Height

Canopy height is considered the key attribute for understanding the vertical structure of forests and is a crucial parameter for modelling forest growth. LiDAR is the best remote-sensing method available for tree height measurement [75]. It has been suggested that the use of LiDAR and other sensors can potentially be a substitute for field measured canopy heights [43].

Spectral reflectance in the form of either raw band value or band ratios (e.g. NDVI) is the primary input from satellite imagery when integrated with LiDAR data to estimate height. Combining LiDAR-derived heights with very high-resolution (VHR) imagery and aerial photos provides the means to more accurately segment individual trees hence to estimate forest attributes at the individual tree level [76•]. McCombs et al. [77] combined small-foot print LiDAR and high-resolution digital multispectral imagery captured by Spectral Visions to identify individual stems and estimate plot mean height in two planted pine forests. A focal search function was used to identify the location of individual trees based on LiDAR-derived tree height and NIR reflectance from an image. Tree identification accuracy increased by 7.6 % for high-density plots and 18.8 % for low-density plots using combined sensors. However, no improvement in height was reported using the combined approach. Suárez et al. [78] used high-resolution aerial photography and LiDAR-derived tree canopy model (TCM) to segment individual Sitka spruce trees in order to derive individual tree height. The study found the method successfully estimated individual tree height, especially for larger trees (R 2 = 0.86). The study only used the combined sensor approach; hence, no comparison with a single sensor was reported. Popescu and Wynne [79] utilised multispectral ATLAS imagery to stratify forest types using maximum likelihood classification, which facilitated the identification of single trees and the estimation of their heights with LiDAR. Single trees were identified using a variable window technique with local maximum focal filtering over both LiDAR and ATLAS features. Minor improvement on height estimation with sensor fusion was observed (R 2 increased by 1 % for pines and up to 3 % for deciduous). A canopy fuel study also observed a minor improvement in canopy height estimation using metrics derived from both LiDAR and DMC imagery compared to LiDAR alone (R 2 of 0.957 and 0.935, respectively). The study also examined the canopy base height which is the lowest green foliage and found a 6 % improvement with the sensor fusion [80].

The fusion of LiDAR and optical sensors has been used in both identifying or delineating individual trees and estimating single tree height. There are a variety of statistical algorithms adopted to predict tree height based on metrics derived from LiDAR and optical sensors. All the height estimation studies we reviewed (Table 3) applied a linear or multiple regression equation solved using ordinary least squares (OLS). Model performances were evaluated by the coefficient of determination (R 2) and root mean square error (RMSE).

Table 3 Recent studies of forest height estimation using integrated LiDAR and optical sensors

Table 3 shows that LiDAR has already proven very accurate in canopy height estimation. Fusion with other optical sensors added little improvement in height estimation accuracy (1–7 %). One study even reported a slight negative influence on height estimation with sensor fusion compared to using LiDAR alone [77]. A benefit of sensor fusion for tree height estimation is that it supports tree delineation, which is difficult to achieve with small-footprint or lower point density LiDAR alone. All studies in Table 3 acquired low- to medium-density LiDAR, ranging from 1 to 4 points/m2. The point densities here are lower and have less variation than point densities reported in studies focused on area delineation and species classification. The high R 2 values are evidence that despite the lower point density, tree and canopy height can be estimated with high accuracy.

Forest Volume and Biomass

Height and diameter measurements are highly correlated with forest volume and aboveground biomass in deciduous, coniferous and tropical forests [46]. Volume provides an understanding of forest productivity and structure. Forest biomass allows the estimation of the carbon content in forests, which is important for understanding carbon stock changes associated with deforestation, forest degradation and afforestation. Carbon reporting is required to fulfil obligations to international agreements such as the United Nations Framework Convention on Climate Change (UNFCCC) [81]. Additionally, accurate estimation of biomass in forested areas is essential for developing sustainable climate-friendly strategies [82]. Forest volume and biomass can be directly measured in the field through destructive sampling, but usually, they are estimated based on measured variables such as diameter and height. Remote sensing has been widely applied to estimate forest volume and biomass due to extensive coverage and cost efficiency [50].

Fusion of LiDAR and optical sensors has been studied for estimating forest volume and biomass at the individual tree level [83] or plot level [84] and achieves more accurate estimation than single sensors. An early study used small-footprint airborne LiDAR and multispectral ATLAS imagery to explore the possibility of estimating forest volume and biomass in mixed-species forests. Popescu et al. [85] used spectral information and a LiDAR-derived CHM to delineate individual tree crowns and estimated tree height, basal area, volume and biomass at the plot level using linear regression models. Promising results were achieved for coniferous forests, with R 2 of 0.83 for volume and 0.82 for biomass. Though it was not quantified, the authors commented that an improvement in estimation accuracy was achieved with sensor fusion compared to LiDAR alone. Tonolli et al. [84] estimated tree stem volume at plot level with integration of low-density discrete return airborne LiDAR and multispectral IRS data. Height metrics and canopy variables extracted from LiDAR, together with spectral bands and band ratios extracted from IRS were correlated with the tree volume calculated from field inventory data. The combined sensor approach consistently produced higher R 2 and lower RMSE across all species compared to using either sensor alone. Wallerman and Holmgren [86] predicted forest height, stem density and mean volume using a canonical correlation approach with inputs derived from both LiDAR and SPOT 5 HRG. Using the combined sensors achieved a more accurate mean stand volume estimation compared with using LiDAR only, resulting in a 2–4 % lower RMSE. Estornell et al. [87] used stepwise regression to find the best fit between LiDAR and spectral data from digital images against field measurements, they observed a significant improvement in modelling vegetation volume and biomass when both LiDAR and digital images were used, with R 2 increasing by 12 % for volume and 29 % for biomass compared to LiDAR-only models.

Most studies that we reviewed related to volume and/or biomass estimation using linear or multiple regression models to correlate LiDAR and optical sensor derived metrics with field measurements. However, OLS regression assumes that the explanatory variables are free of measurement errors, which is not realistic in remote sensing data inputs [88]. So, there has been increasing use of non-parametric modelling approaches to model forest volume with integrated LiDAR and spectral inputs. For example, Packalen and Maltamo [89] applied k most similar neighbour (k-MSN) imputation to predict species-specific volume at the stand level from airborne LiDAR-derived height metrics and aerial photograph-derived spectral and textural features. They found that the basal area and volume estimation using inputs from both sensors were accurate (RMSE = 8.63 and 10.36 %, respectively), although deciduous species showed higher errors than coniferous species (up to 34 % higher RMSE). Cartus et al. [90] used a random forest (RF) regression tree algorithm to predict stand-level canopy height and growing stock volume (GSV) over a large plantation area, using metrics derived from low-density discrete return LiDAR, RADAR sensor (ALOS PALSA) and Landsat ETM+. The fusion of RADAR and Landsat produced a minor improvement (6 %) in GSV estimation compared with using LiDAR. Although the study did not combine LiDAR with Landsat, it demonstrated a well-described methodology of using RF to model forests attributes with different sensor inputs.

Overall, optical sensors alone have not shown good performance in assessing forest volume and biomass; recent studies have demonstrated more significant improvement (up to 55 %) in modelling forest volume and biomass with the addition of LiDAR (Table 4). In terms of the measuring model performance, R 2 and RMSE are commonly used. Some studies also measure the bias of models [91]. A wide range of R 2 values was reported for data fusion approaches to estimating both biomass and volume. Biomass estimates ranged from R 2 = 0.33 to 0.82, while volume estimates ranged from R 2 = 0.39 to 0.87. In both cases, the lowest R 2 values were associated with estimates for deciduous forests. It appears that the biomass and volume for deciduous forests are more challenging to predict even with integration of LiDAR and optical sensors. This is an area where the development of species-specific techniques could significantly improve results. Furthermore, most studies reviewed in this section used low- to medium-density discrete return airborne LiDAR. Higher-density LiDAR data or other forms of LiDAR acquisition, such as full waveform or terrestrial LiDAR, could potentially enhance the volume and biomass estimation particularly for deciduous forests due to analysis at a finer scale.

Table 4 Recent studies of forest volume and biomass using integrated LiDAR and optical sensors

Forest Age and Productivity

Too few studies prevent firm conclusions from being drawn about the utility of sensor fusion for forest age and productivity estimation. Nevertheless, a summary of existing studies is presented. Forest age is highly correlated with growth and is a key attribute in forest information systems [92]. Stand age is an important variable but it may not always be available, such as for extensive national forest inventories [93] and uneven-aged natural forests [94]. Stand age data could be collected by field inventory and ring count analysis, but this is labour-intensive and lacks consistency among different cruising crews [95, 96]. Therefore, alternative approaches are desirable and remote-sensing techniques have been useful in estimating forest age. Research has shown that forest age is correlated with spectral reflectance of satellite imagery [97], especially the near infrared band and its derived vegetation indices [98]. Remote-sensing data is also capable of estimating age or age classes based on the differences in tree size, density, understory and canopy developments [99].

Because tree height is a strong predictor of forest age and active sensors directly measure height, studies have incorporated active sensors into forest age estimation to improve accuracy. Vega and St-Onge [100] developed a method of classifying unknown forest age based on existing age-height correlations and a time-series of CHMs derived from both aerial photography and LiDAR. CHMs were reconstructed for a period of 58 years based on historic AP-derived CHMs and a recent LiDAR-derived CHM. The RMSE of the estimated forest age was 7 years.

Forest site productivity refers to the potential tree height or aboveground wood volume for a particular forest site [101]. Understanding patterns in forest productivity is critical to forest resource management and influences tree species selection, design of optimal silvicultural regimes and forecasting timber yields [102]. Site productivity is often expressed as site index (SI), which is calculated as a function of the height of dominant or co-dominant trees at a reference age [101, 103]. SI is a widely accepted quantification of productivity, as stand height or current height growth seems to correlate well with stand volume growth [101]. Also, compared to comprehensive assessments of forest site condition, SI is a simple variable that is relatively easy and inexpensive to measure and is generally not affected by management practices [104]. However, in order to estimate SI, stand age must be known.

The information on tree height at various ages used for building SI is normally derived from measurements of trees from sample plots [105]. However, field measurements sometimes cannot capture the spatial variation of forest productivity due to limited scale [100]. Improvements in availability of remotely sensed data such as LiDAR have yielded opportunities for estimating site productivity at increasingly finer scales. Studies have successfully utilised LiDAR to estimate dominant tree height and evaluate SI and site type for boreal forest in Finland [106], radiata pine in Australia [107] and eucalyptus plantations in Brazil [108] with given forest age. Alternatively, it is also possible to determine SI from single-tree-based LiDAR to extract tree height at the individual tree level so that site index can be predicted at stand level. For example, Gatziolis [109] overlaid field-delineated individual tree crowns on a LiDAR point cloud to estimate dominant tree height; together with stand age information collected from the field, the SI at plot level was estimated. Chen and Zhu [104] used LiDAR-derived heights for individual trees in combination with stand age to predict stand level SI for radiata pine plantations in Australia.

Although LiDAR can accurately estimate tree heights and hence SI, the use of a LiDAR sensor alone to estimate productivity can be limiting, especially if forests lack stand age information. So far, very few studies have used integrated sensors to evaluate site productivity. Lefsky et al. [110] derived stand age by classifying a multi-temporal sequence of Landsat MSS and TM images and extracted stand height and aboveground net primary production of wood (NPP AW) from field measurements and LiDAR transects. In this study NPP AW, which is calculated as the average increment in biomass over a time period, was used as the indicator for forest productivity. A study mentioned earlier by Vega and St-Onge [100] used aerial photos obtained between 1945 and 2003 and recent LiDAR data to reconstruct CHMs over 58 years to estimate SI for jack pine over an extended region; stand SI was estimated with an average RMSE of 2.41 m. The approach developed produced continuous SI and age maps in a spatially explicit way.

To date, too few integrated sensory studies have estimated forest age and/or productivity to prove or disprove the utility of the approach. Most age estimation research today is focussed on using a single sensor or using time-series analysis of remote sensing data [94, 97, 98, 110]. Both productivity studies [100, 110] used optical sensors to gain stand age information in order to evaluate forest productivity, yet neither was considered a sensor fusion approach as the inputs from both LiDAR and optical sensors were not used simultaneously. As a consequence of the lack of previous studies, we are unable to draw conclusions of whether sensor fusion can improve forest age or productivity estimation, but promise has been shown.

Conclusion

The studies reviewed in this paper confirmed that fusion of primarily airborne LiDAR and optical sensors can improve many aspects of forest description. In particular, sensor fusion significantly improved delineation of forest areas (by up to 20 %), identification of species (by up to 21 %), and estimation of forest volume and biomass (by up to 55 %). As LiDAR alone has proven very effective in measuring canopy height, improvements in height estimation due to sensor fusion have been relatively small (between 1 and 7 %). The improvement of integrating LiDAR and optical sensors for forest age and productivity assessment cannot be fully evaluated due to the limited number of studies, yet they provide future research directions for sensor fusion application. Some studies also showed estimating forest height [79] and volume [89] of deciduous forests can be more challenging than coniferous forests possibly due to more complex forest structure and seasonal changes.

The approaches applied for classifying forest type and species, estimating forest structural variables with inputs from LiDAR and spectral data are relatively standardised. However, forest delineation appears to lack a common approach to evaluate performance. It was noted that some studies tend to compare the automated stand delineation results with manual interpretation [58, 59], yet in reality manual results are not always available. Mustonen et al. [53] assessed the variation in mean height, diameter and volume within delineated stands as a basis for evaluating results, which may be an appropriate approach for future studies. Moreover, modelling approaches for predicting forest height and volume commonly used linear or multiple regression analysis or non-parametric approaches such as kNN and random forest. None of these works for all situations; therefore, considerations should be given to selecting the most appropriate modelling approach. Brosofske et al. [91] provided a thorough review on modelling approaches for estimating forest variables, which serves as a reference for selecting regression models.

While fusion appears to be effective in a research environment, a challenge will be to operationalise the research such that forestry companies and governments can implement data fusion for improved forest management. Research on implementation will have to consider how to include data fusion into a standard forest mapping environment and also how to fully utilise the advantages of sensor fusion without incurring substantial extra costs. Additionally, the studies reviewed are primarily discrete return airborne LiDAR, which has been well developed and widely studied. Fusion of optical sensors with other forms of LiDAR such as fullwave form and terrestrial scanners is worth exploring in deriving an enhanced forest description.