Estimating Entropy Of Distributions In Constant Space

ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019)(2019)

引用 24|浏览91
暂无评分
摘要
We consider the task of estimating the entropy of k-ary distributions from samples in the streaming model, where space is limited. Our main contribution is an algorithm that requires O(k log(1/epsilon)(2)/epsilon(3)) samples and a constant O(1) memory words of space and outputs a +/-epsilon estimate of H(p). Without space limitations, the sample complexity has been established as S(k, epsilon) = Theta(k/epsilon log k + log(2) k/epsilon(2)), which is sub-linear in the domain size k, and the current algorithms that achieve optimal sample complexity also require nearly-linear space in k.Our algorithm partitions [0, 1] into intervals and estimates the entropy contribution of probability values in each interval. The intervals are designed to trade off the bias and variance of these estimates.
更多
查看译文
关键词
sample complexity
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要