3260 papers • 126 benchmarks • 313 datasets
In statistical hypothesis testing, a two-sample test is a test performed on the data of two random samples, each independently obtained from a different given population. The purpose of the test is to determine whether the difference between these two populations is statistically significant. The statistics used in two-sample tests can be used to solve many machine learning problems, such as domain adaptation, covariate shift and generative adversarial networks.
(Image credit: Papersgraph)
These leaderboards are used to track progress in two-sample-testing-19
Use these libraries to find two-sample-testing-19 models and implementations
No subtasks available.
It is shown that packing naturally penalizes generators with mode collapse, thereby favoring generator distributions with less mode collapse during the training process, and numerical experiments suggests that packing provides significant improvements in practice as well.
This work proposes a measure of 'sensitivity' and shows empirically that normal samples and adversarial samples have distinguishable sensitivity, and integrates statistical hypothesis testing and model mutation testing to check whether an input sample is likely to be normal or adversarial at runtime by measuring its sensitivity.
A fully automatic way to quantify tumor imaging characteristics using deep learning-based segmentation and test whether these characteristics are predictive of tumor genomic subtypes is proposed.
We introduce hyppo, a unified library for performing multivariate hypothesis testing, including independence, two-sample, and k-sample testing. While many multivariate independence tests have R packages available, the interfaces are inconsistent and most are not available in Python. hyppo includes many state of the art multivariate testing procedures. The package is easy-to-use and is flexible enough to enable future extensions. The documentation and all releases are available at https://hyppo.neurodata.io.
This work forms a method that generates an independent sample via a single feedforward pass through a multilayer perceptron, as in the recently proposed generative adversarial networks, using MMD to learn to generate codes that can then be decoded to produce samples.
MMDAgg significantly outperforms alternative state-of-the-art MMD-based two-sample tests on synthetic data satisfying the Sobolev smoothness assumption, and that, on real-world image data, MMDAgg closely matches the power of tests leveraging the use of models such as neural networks.
This paper develops an efficient online robust PCA method, namely, online moving window robust principal component analysis (OMWRPCA).
This work uses a simple test that takes the mean discrepancy of a witness function as the test statistic and proves that minimizing a squared loss leads to a witness with optimal testing power to achieve competitive performance on a diverse distribution shift benchmark as well as on challenging two-sample testing problems.
This work uses a central limit theorem for the empirical measure in the test statistic of the composite hypothesis Hoeffding test to derive a new estimator for the threshold needed by the test, and finds that this estimator controls better for false alarms while maintaining satisfactory detection probabilities.
Adding a benchmark result helps the community track progress.