The explainability of gradient-boosted decision trees for digital elevation model (dem) error prediction

dc.contributor.authorOkolie, Chukwuma
dc.contributor.authorMills, Jon
dc.contributor.authorAdeleke, Adedayo
dc.contributor.authorSmit, Julian
dc.contributor.authorMaduako, Ikechukwu
dc.contributor.emailadedayo.adeleke@up.ac.zaen_US
dc.date.accessioned2024-05-21T04:58:25Z
dc.date.available2024-05-21T04:58:25Z
dc.date.issued2023-09
dc.description.abstractGradient boosted decision trees (GBDTs) have repeatedly outperformed several machine learning and deep learning algorithms in competitive data science. However, the explainability of GBDT predictions especially with earth observation data is still an open issue requiring more focus by researchers. In this study, we investigate the explainability of Bayesian-optimised GBDT algorithms for modelling and prediction of the vertical error in Copernicus GLO-30 digital elevation model (DEM). Three GBDT algorithms are investigated (extreme gradient boosting - XGBoost, light boosting machine – LightGBM, and categorical boosting – CatBoost), and SHapley Additive exPlanations (SHAP) are adopted for the explainability analysis. The assessment sites are selected from urban/industrial and mountainous landscapes in Cape Town, South Africa. Training datasets are comprised of eleven predictor variables which are known influencers of elevation error: elevation, slope, aspect, surface roughness, topographic position index, terrain ruggedness index, terrain surface texture, vector roughness measure, forest cover, bare ground cover, and urban footprints. The target variable (elevation error) was calculated with respect to accurate airborne LiDAR. After model training and testing, the GBDTs were applied for predicting the elevation error at model implementation sites. The SHAP plots showed varying levels of emphasis on the parameters depending on the land cover and terrain. For example, in the urban area, the influence of vector ruggedness measure surpassed that of first-order derivatives such as slope and aspect. Thus, it is recommended that machine learning modelling procedures and workflows incorporate model explainability to ensure robust interpretation and understanding of model predictions by both technical and non-technical users.en_US
dc.description.departmentGeography, Geoinformatics and Meteorologyen_US
dc.description.sdgSDG-09: Industry, innovation and infrastructureen_US
dc.description.urihttp://www.isprs.org/publications/archives.aspxen_US
dc.identifier.citationOkolie, C., Mills, J., Adeleke, A., Smit, J., and Maduako, I.: The explainability of gradient-boosted decision trees for digital elevation model (dem) error prediction, International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLVIII-M-3-2023, 161–168, https://doi.org/10.5194/isprs-archives-XLVIII-M-3-2023-161-2023, 2023.en_US
dc.identifier.issn1682-1750 (print)
dc.identifier.issn2194-9034 (online)
dc.identifier.other10.5194/isprs-archives-XLVIII-M-3-2023-161-2023
dc.identifier.urihttp://hdl.handle.net/2263/96094
dc.language.isoenen_US
dc.publisherInternational Society of Photogrammetry and Remote Sensingen_US
dc.rights© Author(s) 2023. This work is distributed under the Creative Commons Attribution 4.0 License.en_US
dc.subjectShapley additive explanationsen_US
dc.subjectExtreme gradient boostingen_US
dc.subjectCategorical boostingen_US
dc.subjectLight boosting machineen_US
dc.subjectMachine learning explainabilityen_US
dc.subjectGradient boosted decision trees (GBDTs)en_US
dc.subjectSDG-09: Industry, innovation and infrastructureen_US
dc.titleThe explainability of gradient-boosted decision trees for digital elevation model (dem) error predictionen_US
dc.typeArticleen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Chukwuma_Explainability_2023.pdf
Size:
1.6 MB
Format:
Adobe Portable Document Format
Description:
Article

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: