Go to file
freeneuro 0e5eac09a9
add carla_dataset
2022-04-20 15:40:59 +08:00
.idea add something 2022-04-20 15:33:12 +08:00
assets modify something 2022-03-16 14:21:30 +08:00
src add carla_dataset 2022-04-20 15:40:59 +08:00
README.md Update README.md 2022-03-16 15:01:13 +08:00
_config.yml Set theme jekyll-theme-slate 2021-09-07 16:16:21 +08:00

README.md

[AAAI 2022] FCA: Learning a 3D Full-coverage Vehicle Camouflage for Multi-view Physical Adversarial Attack

Overview

This is the official implementation and case study of the Full-coverage Vehicle Camouflage(FCA) method proposed in our AAAI 2022 paper FCA: Learning a 3D Full-coverage Vehicle Camouflage for Multi-view Physical Adversarial Attack.

Source code can be find in here.

Abstract

Physical adversarial attacks in object detection have attracted increasing attention. However, most previous works focus on hiding the objects from the detector by generating an individual adversarial patch, which only covers the planar part of the vehicles surface and fails to attack the detector in physical scenarios for multi-view, long-distance and partially occluded objects. To bridge the gap between digital attacks and physical attacks, we exploit the full 3D vehicle surface to propose a robust Full-coverage Camouflage Attack (FCA) to fool detectors. Specifically, we first try rendering the nonplanar camouflage texture over the full vehicle surface. To mimic the real-world environment conditions, we then introduce a transformation function to transfer the rendered camouflaged vehicle into a photo realistic scenario. Finally, we design an efficient loss function to optimize the camouflage texture. Experiments show that the full-coverage camouflage attack can not only outperform state-of-the-art methods under various test cases but also generalize to different environments, vehicles, and object detectors. The code of FCA will be available at: https://idrl-lab.github.io/Full-coveragecamouflage adversarial-attack/.

Framework

image-20211209204327675

Cases of Digital Attack

Multi-view Attack: Carmear distance is 3

Elevation 0 Elevation 30 Elevation 50
Original
FCA

Multi-view Attack: Carmear distance is 5

Elevation 20 Elevation 40 Elevation 50
Original
FCA

Multi-view Attack: Carmear distance is 10

Elevation 30 Elevation 40 Elevation 50
Original
FCA

Multi-view Attack: different distance, elevation and azimuth

Original
FCA

Partial occlusion

Original
FCA

Ablation study

Different combination of loss terms

As we can see from the Figure, different loss terms plays different roles in attacking. For example, the camouflaged car generated by obj+smooth (we omit the smooth loss, and denotes as obj) can hidden the vehicle successfully, while the camouflaged car generated by iou can successfully suppress the detecting bounding box of the car region, and finally the camouflaged car generated by cls successfully make the detector to misclassify the car to anther category.

Different initialization ways

original basic initialization random initialization zero initialization

Cases of Phyical Attack

Case in the simulator environment

Case in the real world