Digitala Vetenskapliga Arkivet

Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Finding Minimum-Cost Explanations for Predictions made by Tree Ensembles
Linköping University, Department of Computer and Information Science, Software and Systems. Linköping University, Faculty of Science & Engineering. Saab AB. (RTSLab)ORCID iD: 0000-0002-4073-0417
Linköping University, Department of Mathematics, Applied Mathematics. Linköping University, Faculty of Science & Engineering. Saab AB.ORCID iD: 0000-0002-9498-1924
Linköping University, Department of Computer and Information Science, Software and Systems. Linköping University, Faculty of Science & Engineering.ORCID iD: 0000-0002-1485-0802
(English)Manuscript (preprint) (Other academic)
Abstract [en]

The ability to reliably explain why a machine learning model arrives at a particular prediction is crucial when used as decision support by human operators of critical systems. The provided explanations must be provably correct, and preferably without redundant information, called minimal explanations. In this paper, we aim at finding explanations for predictions made by tree ensembles that are not only minimal, but also minimum with respect to a cost function.

To this end, we first present a highly efficient oracle that can determine the correctness of explanations, surpassing the runtime performance of current state-of-the-art alternatives by several orders of magnitude. 

Secondly, we adapt an algorithm called MARCO from the optimization field (calling the adaptation m-MARCO) to compute a single minimum explanation per prediction, and demonstrate an overall speedup factor of 2.7 compared to a state-of-the-art algorithm based on minimum hitting sets, and a speedup factor of 27 compared to the standard MARCO algorithm which enumerates all minimal explanations.

Keywords [en]
explainable AI, abstract interpretation, tree ensembles, optimization
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:liu:diva-213142DOI: 10.48550/arXiv.2303.09271OAI: oai:DiVA.org:liu-213142DiVA, id: diva2:1953715
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

This is a pre-print publication also published at ArXiv, which has not been under subject to peer-review.

Available from: 2025-04-22 Created: 2025-04-22 Last updated: 2025-04-30
In thesis
1. Efficient Formal Reasoning about the Trustworthiness of Tree Ensembles
Open this publication in new window or tab >>Efficient Formal Reasoning about the Trustworthiness of Tree Ensembles
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Recent advances in machine learning (ML) have demonstrated that statistical models can be trained to recognize complicated patterns and make sophisticated decisions, often outperforming human capabilities. However, current ML-based systems sometimes exhibit faulty behavior, which prevents their use in safety-critical applications. Furthermore, the increased autonomy and complexity of ML-based systems raise additional trustworthiness concerns such as explainability. To mitigate risks associated with these concerns, the European Union has agreed on a legal framework called the AI Act that governs ML-based products that are placed on the European market.

To meet the goals set out by the AI Act with diligence, manufacturers of critical systems that leverage machine learning need new rigorous and scalable approaches to ensure trustworthiness. One natural pathway towards this aim is to leverage formal methods, a class of reasoning approaches that has proved useful in the past when arguing for safety. Unfortunately, these methods often suffer from computational scalability issues, a problem that becomes even more challenging when applied to complex ML-based systems.

This dissertation seeks to assess and improve the trustworthiness of a class of machine learning models called tree ensembles. To this end, a reasoning engine based on abstract interpretation is developed from the ground up, exploiting unique characteristics of tree ensembles for improved runtime performance. The engine is designed to support deductive and abductive reasoning with soundness and completeness guarantees, hence enabling formal verification and the ability to accurately explain why a tree ensemble arrives at a certain prediction.

Through extensive experimentation, we demonstrate speedup improvements of several orders of magnitude compared to current state-of-the-art approaches. More importantly, we show that many classifiers based on tree ensembles are extremely sensitive to additive noise, despite demonstrating high accuracy. For example, in one case study involving the classification of images of handwritten digits, we find that changing the light intensity of a single pixel in an image by as little as 0.1% causes some tree ensembles to misclassify the depicted digit. However, we also show that it is possible to compute provably correct explanations without superfluous information, called minimal explanations, for predictions made by complex tree ensembles. These types of explanations can help to pinpoint exactly which inputs are relevant in a particular classification. Moreover, we explore approaches for computing explanations that are also globally optimal with respect to a cost function, called minimum explanations. These types of explanations can be tailored for specific target audiences, e.g., for engineers trying to improve robustness, for system operators making critical decisions, or for incident investigators seeking the root cause of a hazard.

Place, publisher, year, edition, pages
Linköping: Linköping University Electronic Press, 2025. p. 45
Series
Linköping Studies in Science and Technology. Dissertations, ISSN 0345-7524 ; 2454
National Category
Artificial Intelligence
Identifiers
urn:nbn:se:liu:diva-213383 (URN)10.3384/9789181181296 (DOI)9789181181289 (ISBN)9789181181296 (ISBN)
Public defence
2025-08-25, TEMCAS, TEMA Building, Campus Valla, Linköping, 13:15 (English)
Opponent
Supervisors
Note

Funding agencies: This work was partially supported by the Wallenberg AI, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg Foundation. Some computing resources were provided by the Swedish National Infrastructure for Computing (SNIC) and the Swedish National Supercomputer Centre (NSC).

Available from: 2025-04-30 Created: 2025-04-30 Last updated: 2025-05-05Bibliographically approved

Open Access in DiVA

pre-print publication(1894 kB)28 downloads
File information
File name FULLTEXT01.pdfFile size 1894 kBChecksum SHA-512
489dda1a3eb3e0c8774ffdef779d8b6893d13f6d86e54e12a341c8115ef0024f1c435fe8ca2d405122dd85f52b92106096b3a3489cbe5f7d404e31427a8d40ae
Type fulltextMimetype application/pdf

Other links

Publisher's full text

Search in DiVA

By author/editor
Törnblom, JohnKarlsson, EmilNadjm-Tehrani, Simin
By organisation
Software and SystemsFaculty of Science & EngineeringApplied Mathematics
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 32 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 347 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf