|| Lower bound on Wyner’s Common Information
||Erixhen Sula, Michael Gastpar, École polytechnique fédérale de Lausanne, Switzerland|
||D4-S1-T1: Information Inequalities I
||Thursday, 15 July, 22:00 - 22:20
||Thursday, 15 July, 22:20 - 22:40
An important notion of common information between two random variables is due to Wyner. In this paper, we derive a lower bound on a relaxed variant of Wyner's common information for continuous random variables. The new bound reduces to the lower bound on Wyner's common information of Liu (2018). We also show that the new lower bound is tight for a special case of the so-called ``Gaussian channels", namely, when the joint distribution of the random variables can be written as the sum of a single underlying random variable and Gaussian noises. We motivate this work from the recent variations of Wyner's common information and applications to network data compression problems such as the Gray-Wyner network.