Document detail
ID

oai:arXiv.org:2407.08678

Topic
Computer Science - Machine Learnin... Mathematics - Optimization and Con... Statistics - Computation Statistics - Machine Learning 90C15, 65C35, 68T07
Author
Ding, Zihan Jin, Kexin Latz, Jonas Liu, Chenguang
Category

Computer Science

Year

2024

listing date

7/17/2024

Keywords
machine abram learning
Metrics

Abstract

Deep neural networks and other modern machine learning models are often susceptible to adversarial attacks.

Indeed, an adversary may often be able to change a model's prediction through a small, directed perturbation of the model's input - an issue in safety-critical applications.

Adversarially robust machine learning is usually based on a minmax optimisation problem that minimises the machine learning loss under maximisation-based adversarial attacks.

In this work, we study adversaries that determine their attack using a Bayesian statistical approach rather than maximisation.

The resulting Bayesian adversarial robustness problem is a relaxation of the usual minmax problem.

To solve this problem, we propose Abram - a continuous-time particle system that shall approximate the gradient flow corresponding to the underlying learning problem.

We show that Abram approximates a McKean-Vlasov process and justify the use of Abram by giving assumptions under which the McKean-Vlasov process finds the minimiser of the Bayesian adversarial robustness problem.

We discuss two ways to discretise Abram and show its suitability in benchmark adversarial deep learning experiments.

Ding, Zihan,Jin, Kexin,Latz, Jonas,Liu, Chenguang, 2024, How to beat a Bayesian adversary

Document

Open

Share

Source

Articles recommended by ES/IODE AI