{ "id": "2006.03689", "version": "v1", "published": "2020-06-05T21:05:19.000Z", "updated": "2020-06-05T21:05:19.000Z", "title": "Anomaly Detection with Domain Adaptation", "authors": [ "Ziyi Yang", "Iman Soltani Bozchalooi", "Eric Darve" ], "categories": [ "cs.LG", "stat.ML" ], "abstract": "We study the problem of semi-supervised anomaly detection with domain adaptation. Given a set of normal data from a source domain and a limited amount of normal examples from a target domain, the goal is to have a well-performing anomaly detector in the target domain. We propose the Invariant Representation Anomaly Detection (IRAD) to solve this problem where we first learn to extract a domain-invariant representation. The extraction is achieved by an across-domain encoder trained together with source-specific encoders and generators by adversarial learning. An anomaly detector is then trained using the learnt representations. We evaluate IRAD extensively on digits images datasets (MNIST, USPS and SVHN) and object recognition datasets (Office-Home). Experimental results show that IRAD outperforms baseline models by a wide margin across different datasets. We derive a theoretical lower bound for the joint error that explains the performance decay from overtraining and also an upper bound for the generalization error.", "revisions": [ { "version": "v1", "updated": "2020-06-05T21:05:19.000Z" } ], "analyses": { "keywords": [ "domain adaptation", "target domain", "irad outperforms baseline models", "digits images datasets", "anomaly detector" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }