Universal adversarial perturbations against object detection

highlights•We propose an algorithm to generate universal adversarial perturbations against object detection. To the best of our knowledge, this work is the first one that empirically proves the existence of such perturbations, which can lead the target detector to fail in finding any objects on most...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Pattern recognition 2021-02, Vol.110, p.107584, Article 107584
Hauptverfasser: Li, Debang, Zhang, Junge, Huang, Kaiqi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:highlights•We propose an algorithm to generate universal adversarial perturbations against object detection. To the best of our knowledge, this work is the first one that empirically proves the existence of such perturbations, which can lead the target detector to fail in finding any objects on most images.•We introduce two criteria to evaluate the blind degree of the detectors and show that detectors are highly vulnerable to such universal perturbations.•We analyze the generalization of the universal perturbations across different training sets, backbone networks, and detectors, which shows promising universality in such black-box attack settings.•We also use the proposed method to generate the class-specific universal perturbations, which can remove the detection results of the target class and keep the results of other classes unchanged. Despite the remarkable success of deep neural networks on many visual tasks, they have been proved to be vulnerable to adversarial examples. For visual tasks, adversarial examples are images added with visually imperceptible perturbations that result in failure for recognition. Previous works have demonstrated that adversarial perturbations can cause neural networks to fail on object detection. But these methods focus on generating an adversarial perturbation for a specific image, which is the image-specific perturbation. This paper tries to extend such image-level adversarial perturbations to detector-level, which are universal (image-agnostic) adversarial perturbations. Motivated by this, we propose a Universal Dense Object Suppression (U-DOS) algorithm to derive the universal adversarial perturbations against object detection and show that such perturbations with visual imperceptibility can lead the state-of-the-art detectors to fail in finding any objects in most images. Compared to image-specific perturbations, the results of image-agnostic perturbations are more interesting and also pose more challenges in AI security, because they are more convenient to be applied in the real physical world. We also analyze the generalization of such universal adversarial perturbations across different detectors and datasets under the black-box attack settings, showing it’s a simple but promising adversarial attack approach against object detection. Furthermore, we validate the class-specific universal perturbations, which can remove the detection results of the target class and keep others unchanged.
ISSN:0031-3203
1873-5142
DOI:10.1016/j.patcog.2020.107584