Yifan Jiang

Wasserstein distributional robustness of neural networks
Date
Apr 20, 2024, 2:20 pm2:35 pm

Speakers

Details

Event Description

Deep neural networks are known to be vulnerable to adversarial attacks (AA). We recast AA using techniques of Wasserstein distributionally robust optimization (DRO) and obtain novel contributions leveraging recent insights from DRO sensitivity analysis. We consider a set of distributional threat models which allow attackers to perturb inputs in a non-uniform way. We link these more general attacks with questions of out-of-sample performance and Knightian uncertainty. To evaluate the distributional robustness of neural networks, we propose a first-order AA algorithm and its multi-step version. Furthermore, we provide a new asymptotic estimate of the adversarial accuracy against distributional threat models. The bound is fast to compute and first-order accurate, offering new insights even for the pointwise AA. It also naturally yields out-of-sample performance guarantees. We conduct numerical experiments on the CIFAR-10, CIFAR-100, and ImageNet datasets using DNNs on RobustBench to illustrate our theoretical results.