Connecting the Digital and Physical World: Improving the Robustness of Adversarial Attacks


Abstract
While deep learning models have achieved unprecedented success in various domains, there is also a growing concern of adversarial attacks against related applications. Recent results show that by adding a small amount of perturbations to an image (imperceptible to humans), the resulting adversarial examples can force a classifier to make targeted mistakes. So far, most existing works focus on crafting adversarial examples in the digital domain, while limited efforts have been devoted to understanding the physical domain attacks. In this work, we explore the feasibility of generating robust adversarial examples that remain effective in the physical domain. Our core idea is to use an image-to-image translation network to simulate the digital-to-physical transformation process for generating robust adversarial examples. To validate our method, we conduct a large-scale physical-domain experiment, which involves manually taking more than 3000 physical domain photos. The results show that our method outperforms existing ones by a large margin and demonstrates a high level of robustness and transferability.

System Overview


Material


Bibtex
@inproceedings{Steve_d2p_2019,
  author = {Steve T.K. Jan and Joseph Messou and Yen-Chen Lin and Jia-Bin Huang and Gang Wang},
  booktitle = {The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI)},
  title = {Connecting the Digital and Physical World: Improving the Robustness of
Adversarial Attacks},
  year = {2019}
}