Update README.md

This commit is contained in:
freeneuro 2021-09-12 16:33:38 +08:00 committed by GitHub
parent 59c909da94
commit b5a116dce1
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
1 changed files with 18 additions and 3 deletions

View File

@ -1,4 +1,19 @@
# Full-coverage-camouflage-adversarial-attack ## FCA: Learning a 3D Full-coverage Vehicle Camouflage for Multi-view Physical Adversarial Attack
## code and example are public available!
### Different combination of loss terms Case study of the FCA. The code can be find in [FCA]([http://]).
### Cases of digital attack
!(image)[https://github.com/winterwindwang/Full-coverage-camouflage-adversarial-attack/blob/gh-pages/assets/distance_10_elevation_30_adv_pred.gif]
### Cases of multi-view robust
### Ablation study
#### Different combination of loss terms
![image](https://github.com/winterwindwang/Full-coverage-camouflage-adversarial-attack/blob/gh-pages/assets/abaltion_study_loss.png) ![image](https://github.com/winterwindwang/Full-coverage-camouflage-adversarial-attack/blob/gh-pages/assets/abaltion_study_loss.png)
As we can see from the Figure, different loss term plays different role in attacking. For example, the camouflaged car generated by `obj+smooth (we omit the smooth loss, and denotes as obj)` hardly hidden from the detector, while the camouflaged car generated by `iou` successfully suppress the detecting bounding box of the car region, and finally the camouflaged car generated by `cls` successfully make the detector to misclassify the car to anther category.