Poster
Sharp Optimality of Simple, Plug-in Estimation of the Fisher Information of a Smoothed Density
Subhodh Kotekal
West Exhibition Hall B2-B3 #W-812
It is often difficult to estimate even simple functions of an unknown, data-generating distribution given access only to samples, particularly when the distribution is quite complicated. It is typically suboptimal (in a statistical sense) to estimate by just computing the target function on the empirical distribution of the data points. Classical work has proposed many complicated techniques to do optimal estimation, but the methodologies appear brittle and too tailored. A quite different approach recently proposed in the literature is to add noise to the data, with the idea that the noised distribution will become simpler, looking more and more like the noise distribution. It may now be optimal to simply compute the target on the empirical distribution of the noised data points. In this paper, we sharply characterize the statistical limits of estimating a fundamental functional, the Fisher information, of a smooth density corrupted by additive Gaussian noise. Interestingly, there is a critical level of noise below which the problem is actually harder than if there were no noise. Nevertheless, we show that the simple plug-in strategy is indeed optimal. Using some fundamental identities, we describe how to estimate other information-theoretic quantities, like the entropy and the mutual information, from our results. Our results provide support for the noising approach.