Cargando…
An Upper Bound on the Error Induced by Saddlepoint Approximations—Applications to Information Theory †
This paper introduces an upper bound on the absolute difference between: [Formula: see text] the cumulative distribution function (CDF) of the sum of a finite number of independent and identically distributed random variables with finite absolute third moment; and [Formula: see text] a saddlepoint a...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517223/ https://www.ncbi.nlm.nih.gov/pubmed/33286462 http://dx.doi.org/10.3390/e22060690 |
Sumario: | This paper introduces an upper bound on the absolute difference between: [Formula: see text] the cumulative distribution function (CDF) of the sum of a finite number of independent and identically distributed random variables with finite absolute third moment; and [Formula: see text] a saddlepoint approximation of such CDF. This upper bound, which is particularly precise in the regime of large deviations, is used to study the dependence testing (DT) bound and the meta converse (MC) bound on the decoding error probability (DEP) in point-to-point memoryless channels. Often, these bounds cannot be analytically calculated and thus lower and upper bounds become particularly useful. Within this context, the main results include, respectively, new upper and lower bounds on the DT and MC bounds. A numerical experimentation of these bounds is presented in the case of the binary symmetric channel, the additive white Gaussian noise channel, and the additive symmetric [Formula: see text]-stable noise channel. |
---|