Classifier-Based Nonparametric Sequential Hypothesis Testing

2026-03-20Information Theory

Information Theory
AI summary

The authors study how to make decisions step-by-step about which of several possible sources a stream of data comes from, using past data to train a classifier. They propose a method that can steadily test whether the data belongs to a null group or one of several alternative groups, with guarantees on how quickly it makes decisions and how accurate it is. Their approach also allows correctly identifying the exact source distribution eventually. They explore conditions that ensure the classifier works well, consider how having different training and testing data affects results, and show examples with both simulated and real data.

sequential testingpower-one testmulti-class classifierdistribution separabilityexpected stopping timeoffline datasetchange detectiontraining-testing mismatchhypothesis testing
Authors
Chia-Yu Hsu, Shubhanshu Shekhar
Abstract
We consider the problem of constructing sequential power-one tests where the null and alternative classes are specified indirectly through historical or offline data. More specifically, given an offline dataset consisting of observations from $L+1$ distributions $\{P_0, P_1, \ldots, P_L\}$, and a new unlabeled data stream $\{X_t: t \geq 1\} \overset{i.i.d}{\sim} P_θ$, the goal is to decide between the null $H_0: θ= 0$, against the alternative $H_1: θ\in [L]:=\{1,\ldots,L\}$. Our main methodological contribution is a general approach for designing a level-$α$ power-one test for this problem using a multi-class classifier trained on the given offline dataset. Working under a mild "separability" condition on the distributions and the trained classifier, we obtain an upper bound on the expected stopping time of our proposed level-$α$ test, and then show that in general this cannot be improved. In addition to rejecting the null, we show that our procedure can also identify the true underlying distribution almost surely. We then establish a sufficient condition to ensure the required separability of the classifier, and provide some converse results to investigate the role of the size of the offline dataset and the family of classifiers among classifier-based tests that satisfy the level-$α$ power-one criterion. Finally, we present an extension of our analysis for the training-and-testing distribution mismatch and illustrate an application to sequential change detection. Empirical results using both synthetic and real data provide support for our theoretical results.