All Dates/Times are Australian Eastern Standard Time (AEST)

# Technical Program

## Paper Detail

 Paper ID D5-S3-T3.2 Paper Title Fisher Information and Mutual Information Constraints Authors Leighton Pate Barnes, Ayfer Ozgur, Stanford University, United States Session D5-S3-T3: Statistics Chaired Session: Friday, 16 July, 22:40 - 23:00 Engagement Session: Friday, 16 July, 23:00 - 23:20 Abstract We consider the processing of statistical samples $X\sim P_\theta$ by a channel p(y|x), and characterize how the statistical information from the samples for estimating the parameter $\theta\in\mathbb{R}^d$ can scale with the mutual information or capacity of the channel. We show that if the statistical model has a sub-Gaussian score function, then the trace of the Fisher information matrix for estimating $\theta$ from Y can scale at most linearly with the mutual information between X and Y. We apply this result to obtain minimax lower bounds in distributed statistical estimation problems, and obtain a tight preconstant for Gaussian mean estimation. We then show how our Fisher information bound can also imply mutual information or Jensen-Shannon divergence based distributed strong data processing inequalities.