Categories
Nevin Manimala Statistics

Classification logit two-sample testing by neural networks for differentiating near manifold densities

IEEE Trans Inf Theory. 2022 Oct;68(10):6631-6662. doi: 10.1109/tit.2022.3175691. Epub 2022 May 17.

ABSTRACT

The recent success of generative adversarial networks and variational learning suggests that training a classification network may work well in addressing the classical two-sample problem, which asks to differentiate two densities given finite samples from each one. Network-based methods have the computational advantage that the algorithm scales to large datasets. This paper considers using the classification logit function, which is provided by a trained classification neural network and evaluated on the testing set split of the two datasets, to compute a two-sample statistic. To analyze the approximation and estimation error of the logit function to differentiate near-manifold densities, we introduce a new result of near-manifold integral approximation by neural networks. We then show that the logit function provably differentiates two sub-exponential densities given that the network is sufficiently parametrized, and for on or near manifold densities, the needed network complexity is reduced to only scale with the intrinsic dimensionality. In experiments, the network logit test demonstrates better performance than previous network-based tests using classification accuracy, and also compares favorably to certain kernel maximum mean discrepancy tests on synthetic datasets and hand-written digit datasets.

PMID:37810208 | PMC:PMC10558099 | DOI:10.1109/tit.2022.3175691

By Nevin Manimala

Portfolio Website for Nevin Manimala