ABSTRACT
The paper introduces the explainable artificial intelligence methodology to assess the global feature importance of the machine learning models used for heat demand forecasting in intelligent control of district heating systems, with motivation facilitate their interpretability and trustworthiness, hence addressing the challenges related to adherence to communal standards, customer satisfaction and liability risks. Methodology involves generation of global feature importance insights by using four different approaches, namely intrinsic (ante-hoc) interpretability of gradient boosting method and selected post-hoc methods, namely partial dependence, accumulated local effects and SHAP and qualitative analysis of those insights in context of expected behavior of district heating systems and comparative analysis. None of the selected methods assume feature permutation or perturbations which can introduce bias due to introduction of random unrealistic values of data instances. The accumulated local effects and SHAP have been found as most reliable methods for determining the feature importance, taking into account feature interactions and non-linearities. The accumulated local effects plots with transmitted energy across the range of ambient temperatures closely resemble the shape of the control curve, which is the evidence of accurate model, as well as suitability of explanation method. By providing the insights which align with the domain expertise, the discussion confirms the value of using explainable artificial intelligence stack as mandatory layer in assessing the performance of machine learning models, especially in high risk artificial intelligence systems, such as those whose use is anticipated in the district heating systems.
KEYWORDS
PAPER SUBMITTED: 2024-12-23
PAPER REVISED: 2025-01-23
PAPER ACCEPTED: 2025-01-30
PUBLISHED ONLINE: 2025-04-05
THERMAL SCIENCE YEAR
2025, VOLUME
29, ISSUE
Issue 5, PAGES [3355 - 3365]
- Zdravković, M., et al., Explainable Heat Demand Forecasting for the Novel Control Strategies of District Heating Systems, Annual Reviews in Control, 53 (2022), Apr., pp. 405-418
- Runge, J., Etienne S., A Comparison of Prediction and Forecasting Artificial Intelligence Models to Estimate the Future Energy Demand in a District Heating System, Energy, 269 (2023), 126661
- Wei, Z., et al., Prediction of Residential District Heating Load Based on Machine Learning: A Case Study, Energy, 231 (2021), 120950
- Hochreiter, S., Schmidhuber, J., Long Short-Term Memory, Neural Computation, 9 (1997), 8, pp. 1735-1780
- Gong, M., et al., Gradient Boosting Machine for Predicting Return Temperature of District Heating System: A Case Study for Residential Buildings in Tianjin, Journal of Building Engineering, 27 (2020), 100950
- Zdravković, M., XGBoost "is All You Need", in: The Case of Forecasting Transmitted Heat Energy in District Heating Systems, (Eds. Trajanović, M., Filipović, N., Zdravković, M.), Disruptive Information Technologies for a Smart Society, ICIST 2024. Lecture Notes in Networks and Systems, Springer, Cham, Switzerland, 2024, Vol. 860
- Huang, Y., et al., Explainable district heat load forecasting with active deep learning, Applied Energy, 350 (2013) 121753
- Ribeiro, M. T., et al., Why Should I Trust You, Explaining the Predictions of Any Classifier, Proceedings, 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, Cal., USA, 2016, pp. 1135-1144
- Dang, L. M., et al., Toward Explainable Heat Load Patterns Prediction for District Heating, Scientific Reports, 13 (2023), 1, 7434
- Lundberg, S., Lee, S.-I., A Unified Approach to Interpreting Model Predictions, On-line first, arxiv.org/abs/1705.07874, 2017
- Park, S., et al., Explainable Anomaly Detection for District Heating Based on Shapley Additive Explanations, Proceedings, International Conference on Data Mining Workshops (ICDMW), Sorrento, Italy, 2020
- Friedman, J. H., Greedy Function Approximation: A Gradient Boosting Machine. The Annals of Statistics, 29 (2001), 5, pp. 1189-1232
- Apley, D. W., Zhu, J., Visualizing the Effects of Predictor Variables in Black Box Supervised Learning Models (Version 2), Online first, doi.org/10.48550/arxiv.1612.08468, 2016
- Fisher, A., et al., All Models are Wrong, but Many are Useful: Learning a Variable's Importance by Studying an Entire Class of Prediction Models Simultaneously, On-line first, doi.org/10.48550/arxiv.1801.01489, 2018