Explainable classification by learning human-readable sentences in feature subsets

•Mining of human-readable decision rules from empirical data sets.•Extraction of context-aware human-readable explanations for classification outputs.•Enables human analyst to obtain contextual understanding of decision logic.•Learns population of sentences (sets of inequalities among learned featur...

Full description

Saved in:
Bibliographic Details
Published inInformation sciences Vol. 564; pp. 202 - 219
Main Authors Krishnamurthy, Prashanth, Sarmadi, Alireza, Khorrami, Farshad
Format Journal Article
LanguageEnglish
Published Elsevier Inc 01.07.2021
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:•Mining of human-readable decision rules from empirical data sets.•Extraction of context-aware human-readable explanations for classification outputs.•Enables human analyst to obtain contextual understanding of decision logic.•Learns population of sentences (sets of inequalities among learned feature subsets). We propose a new methodology (Sentences in Feature Subsets, i.e., SiFS) to mine human-readable decision rules from empirical data sets. Unlike opaque classifiers obtained using deep learning, the proposed methodology derives decision rules that are compact and comprised of Boolean logic sentences involving subsets of features in the input data. For this purpose, we develop a new classifier model defined in terms of sets of inequalities among selected features in the input data. To empirically derive suitable inequalities from training data, our approach combines a differentiable representation of sets of Boolean logic sentences, gradient-based optimization of coefficients in the inequalities, a genetic-based algorithm for selection of the subsets of features, and a “goodness” model of sentences to prune and down-select sentences. We present results on synthetic and real-world benchmark datasets to demonstrate efficacy of SiFS in deriving human-readable decision rules. It is seen that SiFS achieves comparable accuracies to the best among various other classification algorithms (accuracies of 95% to 100% on several datasets, F1 scores between 0.95 and 1.0), reasonable computation times (training times of a few seconds for considered datasets), and compact human-readable decision rules (between 1 to 10 sentences of 3 words or less for considered datasets).
ISSN:0020-0255
1872-6291
DOI:10.1016/j.ins.2021.02.031