Beyond Prediction Similarity: ShapGAP for Evaluating Faithful Surrogate Models in XAI

Ettore Mariotti, Adarsa Sivaprasad, Jose Maria Alonso Moral

Research output: Chapter in Book/Report/Conference proceedingPublished conference contribution

7 Citations (Scopus)
1 Downloads (Pure)

Abstract

The growing importance of Explainable Artificial Intelligence (XAI) has highlighted the need to understand the decision-making processes of black-box models. Surrogation, emulating a black-box model (BB) with a white-box model (WB), is crucial in applications where BBs are unavailable due to security or practical concerns. Traditional fidelity measures only evaluate the similarity of the final predictions, which can lead to a significant limitation: considering a WB faithful even when it has the same prediction as the BB but with a completely different rationale. Addressing this limitation is crucial to develop Trustworthy AI practical applications beyond XAI. To address this issue, we introduce ShapGAP, a novel metric that assesses the faithfulness of surrogate models by comparing their reasoning paths, using SHAP explanations as a proxy. We validate the effectiveness of ShapGAP by applying it to real-world datasets from healthcare and finance domains, comparing its performance against traditional fidelity measures. Our results show that ShapGAP enables better understanding and trust in XAI systems, revealing the potential dangers of relying on models with high task accuracy but unfaithful explanations. ShapGAP serves as a valuable tool for identifying faithful surrogate models, paving the way for more reliable and Trustworthy AI applications.
Original languageEnglish
Title of host publicationExplainable Artificial Intelligence
Subtitle of host publicationFirst World Conference, xAI 2023, Lisbon, Portugal, July 26–28, 2023, Proceedings, Part I
EditorsLuca Longo
PublisherSpringer Nature
Pages160-173
Number of pages14
ISBN (Electronic)978-3-031-44064-9
ISBN (Print)978-3-031-44063-2
DOIs
Publication statusPublished - 30 Oct 2023
EventFirst World Conference - Lisbon, Portugal
Duration: 26 Jul 202328 Jul 2023

Publication series

NameCommunications in Computer and Information Science
Volume1901 CCIS
ISSN (Print)1865-0929
ISSN (Electronic)1865-0937

Conference

ConferenceFirst World Conference
Abbreviated titlexAI 2023
Country/TerritoryPortugal
CityLisbon
Period26/07/2328/07/23

Funding

E. Mariotti and A. Sivaprasad are ESRs in the NL4XAI project which has received funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie Grant Agreement No. 860621. In addition, this work is supported by Grant PID2021-123152OB-C21 funded by MCIN/AEI/10.13039/501100011033 and by “ESF Investing in your future”, by Grant TED2021-130295B-C33 funded by MCIN/AEI/10.13039/501100011033 and by the “European Union NextGenerationEU/PRTR”, and by the Galician Ministry of Culture, Education, Professional Training and University (grants ED431G2019/04, ED431C2022/19 co-funded by the European Regional Development Fund, ERDF/FEDER program).

FundersFunder number
European Research Council 860621

    Keywords

    • Black-box
    • Explainable Artificial Intelligence (XAI)
    • Faithfulness
    • Fidelity Measures
    • Interpretability
    • SHAP
    • Surrogate Models
    • White-box

    Fingerprint

    Dive into the research topics of 'Beyond Prediction Similarity: ShapGAP for Evaluating Faithful Surrogate Models in XAI'. Together they form a unique fingerprint.

    Cite this