An Upper Bound on the Error Induced by Saddlepoint Approximations -- Applications to Information Theory

This paper introduces an upper bound on the absolute difference between: (a) the cumulative distribution function (CDF) of the sum of a finite number of independent and identically distributed random variables with finite absolute third moment; and (b) a saddlepoint approximation of such CDF. This u...

Full description

Saved in:
Bibliographic Details
Published inarXiv.org
Main Authors Anade, Dadja, Gorce, Jean-Marie, Philippe, Mary, Perlaza, Samir
Format Paper Journal Article
LanguageEnglish
Published Ithaca Cornell University Library, arXiv.org 15.07.2020
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:This paper introduces an upper bound on the absolute difference between: (a) the cumulative distribution function (CDF) of the sum of a finite number of independent and identically distributed random variables with finite absolute third moment; and (b) a saddlepoint approximation of such CDF. This upper bound, which is particularly precise in the regime of large deviations, is used to study the dependence testing (DT) bound and the meta converse (MC) bound on the decoding error probability (DEP) in point-to-point memoryless channels. Often, these bounds cannot be analytically calculated and thus lower and upper bounds become particularly useful. Within this context, the main results include, respectively, new upper and lower bounds on the DT and MC bounds. A numerical experimentation of these bounds is presented in the case of the binary symmetric channel, the additive white Gaussian noise channel, and the additive symmetric \(\alpha\)-stable noise channel.
ISSN:2331-8422
DOI:10.48550/arxiv.2007.05319