Evaluating the adversarial robustness of deep networks to gradient-based attacks is challenging. While most attacks consider ω2- and ω→-norm constraints to craft input perturbations, only a few investigate sparse ω1- and ω0-norm attacks. In particular, ω0-norm attacks remain the least studied due to the inherent complexity of optimizing over a non-convex and non-differentiable constraint. However, evaluating adversarial robustness under these attacks could reveal weaknesses otherwise left untested with more conventional ω2- and ω→-norm attacks. In this work, we propose a novel ω0-norm attack, called ε-zero, which leverages a differentiable approximation of the ω0 norm to facilitate gradient-based optimization, and an adaptive projection operator to dynamically adjust the trade-off between loss minimization and perturbation sparsity. Extensive evaluations using MNIST, CIFAR10, and ImageNet datasets, involving robust and non-robust models, show that ε-zero finds minimum ω0-norm adversarial examples without requiring any time-consuming hyperparameter tuning, and that it outperforms all competing sparse attacks in terms of success rate, perturbation size, and efficiency.

\textbackslash sigma-zero: Gradient-based Optimization of \textbackslash ell\_0-norm Adversarial Examples

Antonio Emanuele Cinà
Primo
;
Maura Pintor;Battista Biggio;
2025-01-01

Abstract

Evaluating the adversarial robustness of deep networks to gradient-based attacks is challenging. While most attacks consider ω2- and ω→-norm constraints to craft input perturbations, only a few investigate sparse ω1- and ω0-norm attacks. In particular, ω0-norm attacks remain the least studied due to the inherent complexity of optimizing over a non-convex and non-differentiable constraint. However, evaluating adversarial robustness under these attacks could reveal weaknesses otherwise left untested with more conventional ω2- and ω→-norm attacks. In this work, we propose a novel ω0-norm attack, called ε-zero, which leverages a differentiable approximation of the ω0 norm to facilitate gradient-based optimization, and an adaptive projection operator to dynamically adjust the trade-off between loss minimization and perturbation sparsity. Extensive evaluations using MNIST, CIFAR10, and ImageNet datasets, involving robust and non-robust models, show that ε-zero finds minimum ω0-norm adversarial examples without requiring any time-consuming hyperparameter tuning, and that it outperforms all competing sparse attacks in terms of success rate, perturbation size, and efficiency.
File in questo prodotto:
File Dimensione Formato  
ICLR-2025-sigma-zero-gradient-based-optimization-of-ell_0-norm-adversarial-examples-Paper-Conference.pdf

accesso aperto

Descrizione: versione pubblicata
Tipologia: versione post-print (AAM)
Dimensione 1.35 MB
Formato Adobe PDF
1.35 MB Adobe PDF Visualizza/Apri
appendix_a.pdf

accesso aperto

Descrizione: supplemental
Tipologia: altro documento allegato
Dimensione 7.23 MB
Formato Adobe PDF
7.23 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11584/448845
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact