Repository landing page

We are not able to resolve this OAI Identifier to the repository landing page. If you are the repository manager for this record, please head to the Dashboard and adjust the settings.

A Broad Spectrum Defense Against Adversarial Examples

Abstract

Machine learning models are increasingly employed in making critical decisions across a wide array of applications. As our dependence on these models increases, it is vital to recognize their vulnerability to malicious attacks from determined adversaries. In response to these adversarial attacks, new defensive mechanisms have been developed to ensure the security of machine learning models and the accuracy of the decisions they make. However, many of these mechanisms are reactionary, designed to defend specific models against a known specific attack or family of attacks. This reactionary approach does not generalize to future yet to be developed attacks. In this work, we developed Broad Spectrum Defense (BSD) as a defensive mechanism to secure any model against a wide range of attacks. BSD is not reactionary, and unlike most other approaches, it does not train its detectors using adversarial data, hence removing an inherent bias present in other defenses that rely on having access to adversarial data. An extensive set of experiments showed that BSD outperforms existing detector-based methods such as MagNet and Feature Squeezing. We believe BSD will inspire a new direction in adversarial machine learning to create a robust defense capable of generalizing to existing and future attacks

Similar works

This paper was published in Rowan University.

Having an issue?

Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.