z-logo
open-access-imgOpen Access
Adversarial Attack on Machine Learning Models
Publication year - 2019
Publication title -
international journal of innovative technology and exploring engineering
Language(s) - English
Resource type - Journals
ISSN - 2278-3075
DOI - 10.35940/ijitee.f1088.0486s419
Subject(s) - adversarial system , malware , computer science , adversarial machine learning , artificial intelligence , machine learning , variety (cybernetics) , class (philosophy) , intrusion , computer security , geochemistry , geology
Machine Learning (ML) models are applied in a variety of tasks such as network intrusion detection or malware classification. Yet, these models are vulnerable to a class of malicious inputs known as adversarial examples. These are slightly perturbed inputs that are classified incorrectly by the ML model. The mitigation of these adversarial inputs remains an open problem. As a step towards understanding adversarial examples, we show that they are not drawn from the same distribution than the original data, and can thus be detected using statistical tests. Using this knowledge, we introduce a complimentary approach to identify specific inputs that are adversarial. Specifically, we augment our ML model with an additional output, in which the model is trained to classify all adversarial inputs.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here