CoRTX: Contrastive Framework for Real-time Explanation
Recent advancements in explainable machine learning provide effective and faithful solutions for interpreting model behaviors. However, many explanation methods encounter efficiency issues, which largely limit their deployments in practical scenarios. Real-time explainer (RTX) frameworks have thus b...
Saved in:
Main Authors | , , , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
05.03.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Recent advancements in explainable machine learning provide effective and
faithful solutions for interpreting model behaviors. However, many explanation
methods encounter efficiency issues, which largely limit their deployments in
practical scenarios. Real-time explainer (RTX) frameworks have thus been
proposed to accelerate the model explanation process by learning a
one-feed-forward explainer. Existing RTX frameworks typically build the
explainer under the supervised learning paradigm, which requires large amounts
of explanation labels as the ground truth. Considering that accurate
explanation labels are usually hard to obtain due to constrained computational
resources and limited human efforts, effective explainer training is still
challenging in practice. In this work, we propose a COntrastive Real-Time
eXplanation (CoRTX) framework to learn the explanation-oriented representation
and relieve the intensive dependence of explainer training on explanation
labels. Specifically, we design a synthetic strategy to select positive and
negative instances for the learning of explanation. Theoretical analysis show
that our selection strategy can benefit the contrastive learning process on
explanation tasks. Experimental results on three real-world datasets further
demonstrate the efficiency and efficacy of our proposed CoRTX framework. |
---|---|
DOI: | 10.48550/arxiv.2303.02794 |