Physical adversarial attacks by projecting perturbations

N Worzyk, H Kahlen, O Kramer - … 17–19, 2019, Proceedings, Part III 28, 2019 - Springer
Research on adversarial attacks analyses how to slightly manipulate patterns like images to
make a classifier believe it recogises a pattern with a wrong label, although the correct label …

Edgefool: an adversarial image enhancement filter

AS Shamsabadi, C Oh… - ICASSP 2020-2020 IEEE …, 2020 - ieeexplore.ieee.org
Adversarial examples are intentionally perturbed images that mislead classifiers. These
images can, however, be easily detected using denoising algorithms, when high-frequency …

An analysis of adversarial attacks and defenses on autonomous driving models

Y Deng, X Zheng, T Zhang, C Chen… - 2020 IEEE …, 2020 - ieeexplore.ieee.org
Nowadays, autonomous driving has attracted much attention from both industry and
academia. Convolutional neural network (CNN) is a key component in autonomous driving …

Adaptive wiener filter and natural noise to eliminate adversarial perturbation

F Wu, W Yang, L Xiao, J Zhu - Electronics, 2020 - mdpi.com
Deep neural network has been widely used in pattern recognition and speech processing,
but its vulnerability to adversarial attacks also proverbially demonstrated. These attacks …

Towards cross-task universal perturbation against black-box object detectors in autonomous driving

Q Zhang, Y Zhao, Y Wang, T Baker, J Zhang, J Hu - Computer Networks, 2020 - Elsevier
Deep neural network is the main research branch in artificial intelligence and suitable for
many decision-making fields. Autonomous driving and unmanned vehicle often depend on …

Adversarial example defense based on image reconstruction

H Xu, C Pei, G Yang - PeerJ Computer Science, 2021 - peerj.com
The rapid development of deep neural networks (DNN) has promoted the widespread
application of image recognition, natural language processing, and autonomous driving …

Benchmarking the physical-world adversarial robustness of vehicle detection

T Zhang, Y Xiao, X Zhang, H Li, L Wang - arXiv preprint arXiv:2304.05098, 2023 - arxiv.org
Adversarial attacks in the physical world can harm the robustness of detection models.
Evaluating the robustness of detection models in the physical world can be challenging due …

Principal component adversarial example

Y Zhang, X Tian, Y Li, X Wang… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
Despite having achieved excellent performance on various tasks, deep neural networks
have been shown to be susceptible to adversarial examples, ie, visual inputs crafted with …

Apricot: A dataset of physical adversarial attacks on object detection

A Braunegg, A Chakraborty, M Krumdick… - Computer Vision–ECCV …, 2020 - Springer
Physical adversarial attacks threaten to fool object detection systems, but reproducible
research on the real-world effectiveness of physical patches and how to defend against …

Counteracting adversarial attacks in autonomous driving

Q Sun, AA Rao, X Yao, B Yu, S Hu - Proceedings of the 39th International …, 2020 - dl.acm.org
In this paper, we focus on studying robust deep stereo vision of autonomous driving systems
and counteracting adversarial attacks against it. Autonomous system operation requires real …