A survey of adversarial defenses and robustness in nlp

S Goyal, S Doddapaneni, MM Khapra… - ACM Computing …, 2023 - dl.acm.org
In the past few years, it has become increasingly evident that deep neural networks are not
resilient enough to withstand adversarial perturbations in input data, leaving them …

A survey of safety and trustworthiness of large language models through the lens of verification and validation

X Huang, W Ruan, W Huang, G Jin, Y Dong… - Artificial Intelligence …, 2024 - Springer
Large language models (LLMs) have exploded a new heatwave of AI for their ability to
engage end-users in human-level conversations with detailed and articulate answers across …

Sok: Certified robustness for deep neural networks

L Li, T Xie, B Li - 2023 IEEE symposium on security and privacy …, 2023 - ieeexplore.ieee.org
Great advances in deep neural networks (DNNs) have led to state-of-the-art performance on
a wide range of tasks. However, recent studies have shown that DNNs are vulnerable to …

“real attackers don't compute gradients”: bridging the gap between adversarial ml research and practice

G Apruzzese, HS Anderson, S Dambra… - … IEEE Conference on …, 2023 - ieeexplore.ieee.org
Recent years have seen a proliferation of research on adversarial machine learning.
Numerous papers demonstrate powerful algorithmic attacks against a wide variety of …

Safe planning in dynamic environments using conformal prediction

L Lindemann, M Cleaveland, G Shim… - IEEE Robotics and …, 2023 - ieeexplore.ieee.org
We propose a framework for planning in unknown dynamic environments with probabilistic
safety guarantees using conformal prediction. Particularly, we design a model predictive …

Neuronfair: Interpretable white-box fairness testing through biased neuron identification

H Zheng, Z Chen, T Du, X Zhang, Y Cheng… - Proceedings of the 44th …, 2022 - dl.acm.org
Deep neural networks (DNNs) have demonstrated their outperformance in various domains.
However, it raises a social concern whether DNNs can produce reliable and fair decisions …

" Is your explanation stable?" A Robustness Evaluation Framework for Feature Attribution

Y Gan, Y Mao, X Zhang, S Ji, Y Pu, M Han… - Proceedings of the …, 2022 - dl.acm.org
Neural networks have become increasingly popular. Nevertheless, understanding their
decision process turns out to be complicated. One vital method to explain a models' …

Efficient query-based attack against ML-based Android malware detection under zero knowledge setting

P He, Y Xia, X Zhang, S Ji - Proceedings of the 2023 ACM SIGSAC …, 2023 - dl.acm.org
The widespread adoption of the Android operating system has made malicious Android
applications an appealing target for attackers. Machine learning-based (ML-based) Android …

" Get in Researchers; We're Measuring Reproducibility": A Reproducibility Study of Machine Learning Papers in Tier 1 Security Conferences

D Olszewski, A Lu, C Stillman, K Warren… - Proceedings of the …, 2023 - dl.acm.org
Reproducibility is crucial to the advancement of science; it strengthens confidence in
seemingly contradictory results and expands the boundaries of known discoveries …

Formal verification for neural networks with general nonlinearities via branch-and-bound

Z Shi, Q Jin, JZ Kolter, S Jana, CJ Hsieh, H Zhang - 2023 - openreview.net
Bound propagation with branch-and-bound (BaB) is so far among the most effective
methods for neural network (NN) verification. However, existing works with BaB have mostly …