Cargando…

An Upper Bound on the Error Induced by Saddlepoint Approximations—Applications to Information Theory †

This paper introduces an upper bound on the absolute difference between: [Formula: see text] the cumulative distribution function (CDF) of the sum of a finite number of independent and identically distributed random variables with finite absolute third moment; and [Formula: see text] a saddlepoint a...

Descripción completa

Detalles Bibliográficos
Autores principales: Anade, Dadja, Gorce, Jean-Marie, Mary, Philippe, Perlaza, Samir M.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517223/
https://www.ncbi.nlm.nih.gov/pubmed/33286462
http://dx.doi.org/10.3390/e22060690
Descripción
Sumario:This paper introduces an upper bound on the absolute difference between: [Formula: see text] the cumulative distribution function (CDF) of the sum of a finite number of independent and identically distributed random variables with finite absolute third moment; and [Formula: see text] a saddlepoint approximation of such CDF. This upper bound, which is particularly precise in the regime of large deviations, is used to study the dependence testing (DT) bound and the meta converse (MC) bound on the decoding error probability (DEP) in point-to-point memoryless channels. Often, these bounds cannot be analytically calculated and thus lower and upper bounds become particularly useful. Within this context, the main results include, respectively, new upper and lower bounds on the DT and MC bounds. A numerical experimentation of these bounds is presented in the case of the binary symmetric channel, the additive white Gaussian noise channel, and the additive symmetric [Formula: see text]-stable noise channel.