Categories
Nevin Manimala Statistics

Domain Neural Adaptation

IEEE Trans Neural Netw Learn Syst. 2022 Mar 8;PP. doi: 10.1109/TNNLS.2022.3151683. Online ahead of print.

ABSTRACT

Domain adaptation is concerned with the problem of generalizing a classification model to a target domain with little or no labeled data, by leveraging the abundant labeled data from a related source domain. The source and target domains possess different joint probability distributions, making it challenging for model generalization. In this article, we introduce domain neural adaptation (DNA): an approach that exploits nonlinear deep neural network to 1) match the source and target joint distributions in the network activation space and 2) learn the classifier in an end-to-end manner. Specifically, we employ the relative chi-square divergence to compare the two joint distributions, and show that the divergence can be estimated via seeking the maximal value of a quadratic functional over the reproducing kernel hilbert space. The analytic solution to this maximization problem enables us to explicitly express the divergence estimate as a function of the neural network mapping. We optimize the network parameters to minimize the estimated joint distribution divergence and the classification loss, yielding a classification model that generalizes well to the target domain. Empirical results on several visual datasets demonstrate that our solution is statistically better than its competitors.

PMID:35259116 | DOI:10.1109/TNNLS.2022.3151683

By Nevin Manimala

Portfolio Website for Nevin Manimala