Skip to main content
Log in

On coupling of probability distributions and estimating the divergence through variation

  • Information Theory
  • Published:
Problems of Information Transmission Aims and scope Submit manuscript

Abstract

Let X be a discrete random variable with a given probability distribution. For any α, 0 ≤ α ≤ 1, we obtain precise values for both the maximum and minimum variational distance between X and another random variable Y under which an α-coupling of these random variables is possible. We also give the maximum and minimum values for couplings of X and Y provided that the variational distance between these random variables is fixed. As a consequence, we obtain a new lower bound on the divergence through variational distance.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Zhang, Z., Estimating Mutual Information via Kolmogorov Distance, IEEE Trans. Inform. Theory, 2007, vol. 53, no. 9, pp. 3280–3282.

    Article  MATH  MathSciNet  Google Scholar 

  2. Sason, I., Entropy Bounds for Discrete Random Variables via Maximal Coupling, IEEE Trans. Inform. Theory, 2013, vol. 59, no. 11, pp. 7118–7131.

    Article  MATH  MathSciNet  Google Scholar 

  3. Prelov, V.V., Coupling of Probability Distributions and an Extremal Problem for the Divergence, Probl. Peredachi Inf., 2015, vol. 51, no. 2, pp. 114–121 [Probl. Inf. Trans. (Engl. Transl.), 2015, vol. 51, no. 2, pp. 192–199].

    MATH  MathSciNet  Google Scholar 

  4. Fedotov, A.A., Harremöes, P., and Topsøe, F., Refinements of Pinsker’s Inequality, IEEE Trans. Inform. Theory, 2003, vol. 49, no. 6, pp. 1491–1498.

    Article  MATH  MathSciNet  Google Scholar 

  5. Csiszár, I. and Körner, J., Information Theory: Coding Theorems for Discrete Memoryless Systems, New York: Academic; Budapest: Akad. Kiadó, 1981. Translated under the title Teoriya informatsii: teoremy kodirovaniya dlya diskretnykh sistem bez pamyati, Moscow: Mir, 1985.

    MATH  Google Scholar 

  6. Vajda, I., Note on Discrimination Information and Variation, IEEE Trans. Inform. Theory, 1970, vol. 16, no. 6, pp. 771–773.

    Article  MATH  MathSciNet  Google Scholar 

  7. Ordentlich, E. and Weinberger, M.J., A Distribution Dependent Refinement of Pinsker’s Inequality, IEEE Trans. Inform. Theory, 2005, vol. 51, no. 5, pp. 1836–1840.

    Article  MATH  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to V. V. Prelov.

Additional information

The research was carried out at the Institute for Information Transmission Problems of the Russian Academy of Sciences at the expense of the Russian Science Foundation, project no. 14-50-00150.

Original Russian Text © V.V. Prelov, 2017, published in Problemy Peredachi Informatsii, 2017, Vol. 53, No. 3, pp. 16–22.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Prelov, V.V. On coupling of probability distributions and estimating the divergence through variation. Probl Inf Transm 53, 215–221 (2017). https://doi.org/10.1134/S0032946017030024

Download citation

  • Received:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1134/S0032946017030024

Navigation