The explainability of gradient-boosted decision trees for digital elevation model (dem) error prediction

Show simple item record

dc.contributor.author Okolie, Chukwuma
dc.contributor.author Mills, Jon
dc.contributor.author Adeleke, Adedayo
dc.contributor.author Smit, Julian
dc.contributor.author Maduako, Ikechukwu
dc.date.accessioned 2024-05-21T04:58:25Z
dc.date.available 2024-05-21T04:58:25Z
dc.date.issued 2023-09
dc.description.abstract Gradient boosted decision trees (GBDTs) have repeatedly outperformed several machine learning and deep learning algorithms in competitive data science. However, the explainability of GBDT predictions especially with earth observation data is still an open issue requiring more focus by researchers. In this study, we investigate the explainability of Bayesian-optimised GBDT algorithms for modelling and prediction of the vertical error in Copernicus GLO-30 digital elevation model (DEM). Three GBDT algorithms are investigated (extreme gradient boosting - XGBoost, light boosting machine – LightGBM, and categorical boosting – CatBoost), and SHapley Additive exPlanations (SHAP) are adopted for the explainability analysis. The assessment sites are selected from urban/industrial and mountainous landscapes in Cape Town, South Africa. Training datasets are comprised of eleven predictor variables which are known influencers of elevation error: elevation, slope, aspect, surface roughness, topographic position index, terrain ruggedness index, terrain surface texture, vector roughness measure, forest cover, bare ground cover, and urban footprints. The target variable (elevation error) was calculated with respect to accurate airborne LiDAR. After model training and testing, the GBDTs were applied for predicting the elevation error at model implementation sites. The SHAP plots showed varying levels of emphasis on the parameters depending on the land cover and terrain. For example, in the urban area, the influence of vector ruggedness measure surpassed that of first-order derivatives such as slope and aspect. Thus, it is recommended that machine learning modelling procedures and workflows incorporate model explainability to ensure robust interpretation and understanding of model predictions by both technical and non-technical users. en_US
dc.description.department Geography, Geoinformatics and Meteorology en_US
dc.description.sdg SDG-09: Industry, innovation and infrastructure en_US
dc.description.uri http://www.isprs.org/publications/archives.aspx en_US
dc.identifier.citation Okolie, C., Mills, J., Adeleke, A., Smit, J., and Maduako, I.: The explainability of gradient-boosted decision trees for digital elevation model (dem) error prediction, International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLVIII-M-3-2023, 161–168, https://doi.org/10.5194/isprs-archives-XLVIII-M-3-2023-161-2023, 2023. en_US
dc.identifier.issn 1682-1750 (print)
dc.identifier.issn 2194-9034 (online)
dc.identifier.other 10.5194/isprs-archives-XLVIII-M-3-2023-161-2023
dc.identifier.uri http://hdl.handle.net/2263/96094
dc.language.iso en en_US
dc.publisher International Society of Photogrammetry and Remote Sensing en_US
dc.rights © Author(s) 2023. This work is distributed under the Creative Commons Attribution 4.0 License. en_US
dc.subject Shapley additive explanations en_US
dc.subject Extreme gradient boosting en_US
dc.subject Categorical boosting en_US
dc.subject Light boosting machine en_US
dc.subject Machine learning explainability en_US
dc.subject Gradient boosted decision trees (GBDTs) en_US
dc.subject SDG-09: Industry, innovation and infrastructure en_US
dc.title The explainability of gradient-boosted decision trees for digital elevation model (dem) error prediction en_US
dc.type Article en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record