On Data Classification Efficiency Based on a Trade-off Relation between Mutual Information and Error Probability

We propose a data classification model which yields an average mutual information between a set of objects and a set of class-label decisions as a function of error probability. Optimization of the model consists in minimization of the average mutual information by conditional distributions for the...

Full description

Saved in:
Bibliographic Details
Published in2020 International Conference on Information Technology and Nanotechnology (ITNT) pp. 1 - 6
Main Authors Lange, Mikhail, Lange, Andrey, Paramonov, Semion
Format Conference Proceeding
LanguageEnglish
Published IEEE 26.05.2020
Subjects
Online AccessGet full text
DOI10.1109/ITNT49337.2020.9253225

Cover

More Information
Summary:We propose a data classification model which yields an average mutual information between a set of objects and a set of class-label decisions as a function of error probability. Optimization of the model consists in minimization of the average mutual information by conditional distributions for the decisions subject to a given constraint on the average error probability. It is equivalent to calculating the rate-distortion function in a scheme of coding the source class labels with a given fidelity when a set of the class labels and a set of the objects are connected by an observation channel with known class-conditional probability distributions. Given set of the objects and known observation channel, a lower bound to the rate-distortion function is calculated. This bound is independent on a decision algorithm and yields a potentially achievable error probability subject to a fixed value of the average mutual information. The obtained bound is useful for evaluating an error probability redundancy of any decision algorithm with given discriminant functions.
DOI:10.1109/ITNT49337.2020.9253225