Skip to yearly menu bar Skip to main content


Poster

Optimal Algorithms for Augmented Testing of Discrete Distributions

Maryam Aliakbarpour · Piotr Indyk · Ronitt Rubinfeld · Sandeep Silwal

[ ]
Thu 12 Dec 11 a.m. PST — 2 p.m. PST

Abstract: We consider the problem of hypothesis testing for discrete distributions. In the standard model, where we have sample access to an underlying distribution $p$, extensive research has established optimal bounds for identity testing (goodness of fit), closeness testing (equivalence or two-sample testing), and uniformity testing. We explore these problems in a setting where a predicted data distribution, possibly derived from historical data or predictive machine learning models, is available. We demonstrate that such a predictor can indeed reduce the number of samples required for all three property testing tasks. The reduction in sample complexity depends directly on the predictor’s quality, measured by its total variation distance from $p$. A key advantage of our algorithms is their adaptability to the precision of the prediction. Specifically, our algorithms can self-adjust their sample complexity based on the accuracy of the available prediction, operating without any any prior knowledge of the estimation’s accuracy (i.e. they are consistent). Additionally, we never use more samples than the standard approaches require, even if the predictions provide no meaningful information (i.e. they are also robust). We provide lower bounds to indicate that the improvements in sample complexity achieved by our algorithms are information-theoretically optimal. Furthermore, experimental results show that the performance of our algorithms on real data significantly exceeds our worst-case guarantees for sample complexity, demonstrating the practicality of our approach.

Live content is unavailable. Log in and register to view live content