Evaluating the adversarial robustness of deep networks to gradient-based attacks is challenging. While most attacks consider ω2- and ω→-norm constraints to craft input perturbations, only a few investigate sparse ω1- and ω0-norm attacks. In particular, ω0-norm attacks remain the least studied due to the inherent complexity of optimizing over a non-convex and non-differentiable constraint. However, evaluating adversarial robustness under these attacks could reveal weaknesses otherwise left untested with more conventional ω2- and ω→-norm attacks. In this work, we propose a novel ω0-norm attack, called ε-zero, which leverages a differentiable approximation of the ω0 norm to facilitate gradient-based optimization, and an adaptive projection operator to dynamically adjust the trade-off between loss minimization and perturbation sparsity. Extensive evaluations using MNIST, CIFAR10, and ImageNet datasets, involving robust and non-robust models, show that ε-zero finds minimum ω0-norm adversarial examples without requiring any time-consuming hyperparameter tuning, and that it outperforms all competing sparse attacks in terms of success rate, perturbation size, and efficiency.
\textbackslash sigma-zero: Gradient-based Optimization of \textbackslash ell\_0-norm Adversarial Examples
Antonio Emanuele Cinà
Primo
;Maura Pintor;Battista Biggio;
2025-01-01
Abstract
Evaluating the adversarial robustness of deep networks to gradient-based attacks is challenging. While most attacks consider ω2- and ω→-norm constraints to craft input perturbations, only a few investigate sparse ω1- and ω0-norm attacks. In particular, ω0-norm attacks remain the least studied due to the inherent complexity of optimizing over a non-convex and non-differentiable constraint. However, evaluating adversarial robustness under these attacks could reveal weaknesses otherwise left untested with more conventional ω2- and ω→-norm attacks. In this work, we propose a novel ω0-norm attack, called ε-zero, which leverages a differentiable approximation of the ω0 norm to facilitate gradient-based optimization, and an adaptive projection operator to dynamically adjust the trade-off between loss minimization and perturbation sparsity. Extensive evaluations using MNIST, CIFAR10, and ImageNet datasets, involving robust and non-robust models, show that ε-zero finds minimum ω0-norm adversarial examples without requiring any time-consuming hyperparameter tuning, and that it outperforms all competing sparse attacks in terms of success rate, perturbation size, and efficiency.| File | Dimensione | Formato | |
|---|---|---|---|
|
ICLR-2025-sigma-zero-gradient-based-optimization-of-ell_0-norm-adversarial-examples-Paper-Conference.pdf
accesso aperto
Descrizione: versione pubblicata
Tipologia:
versione post-print (AAM)
Dimensione
1.35 MB
Formato
Adobe PDF
|
1.35 MB | Adobe PDF | Visualizza/Apri |
|
appendix_a.pdf
accesso aperto
Descrizione: supplemental
Tipologia:
altro documento allegato
Dimensione
7.23 MB
Formato
Adobe PDF
|
7.23 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


