Safe learning in robotics: From learning-based control to safe reinforcement learning

L Brunke, M Greeff, AW Hall, Z Yuan… - Annual Review of …, 2022 - annualreviews.org
The last half decade has seen a steep rise in the number of contributions on safe learning
methods for real-world robotic deployments from both the control and reinforcement learning …

Adversarial learning targeting deep neural network classification: A comprehensive review of defenses against attacks

DJ Miller, Z Xiang, G Kesidis - Proceedings of the IEEE, 2020 - ieeexplore.ieee.org
With wide deployment of machine learning (ML)-based systems for a variety of applications
including medical, military, automotive, genomic, multimedia, and social networking, there is …

Safe control with learned certificates: A survey of neural lyapunov, barrier, and contraction methods for robotics and control

C Dawson, S Gao, C Fan - IEEE Transactions on Robotics, 2023 - ieeexplore.ieee.org
Learning-enabled control systems have demonstrated impressive empirical performance on
challenging control problems in robotics, but this performance comes at the cost of reduced …

Smoothllm: Defending large language models against jailbreaking attacks

A Robey, E Wong, H Hassani, GJ Pappas - arXiv preprint arXiv …, 2023 - arxiv.org
Despite efforts to align large language models (LLMs) with human values, widely-used
LLMs such as GPT, Llama, Claude, and PaLM are susceptible to jailbreaking attacks …

Rethinking lipschitz neural networks and certified robustness: A boolean function perspective

B Zhang, D Jiang, D He… - Advances in neural …, 2022 - proceedings.neurips.cc
Designing neural networks with bounded Lipschitz constant is a promising way to obtain
certifiably robust classifiers against adversarial examples. However, the relevant progress …

On mean absolute error for deep neural network based vector-to-vector regression

J Qi, J Du, SM Siniscalchi, X Ma… - IEEE Signal Processing …, 2020 - ieeexplore.ieee.org
In this paper, we exploit the properties of mean absolute error (MAE) as a loss function for
the deep neural network (DNN) based vector-to-vector regression. The goal of this work is …

Globally-robust neural networks

K Leino, Z Wang, M Fredrikson - … Conference on Machine …, 2021 - proceedings.mlr.press
The threat of adversarial examples has motivated work on training certifiably robust neural
networks to facilitate efficient verification of local robustness at inference time. We formalize …

Sok: Certified robustness for deep neural networks

L Li, T Xie, B Li - 2023 IEEE symposium on security and privacy …, 2023 - ieeexplore.ieee.org
Great advances in deep neural networks (DNNs) have led to state-of-the-art performance on
a wide range of tasks. However, recent studies have shown that DNNs are vulnerable to …

Machine unlearning of features and labels

A Warnecke, L Pirch, C Wressnegger… - arXiv preprint arXiv …, 2021 - arxiv.org
Removing information from a machine learning model is a non-trivial task that requires to
partially revert the training process. This task is unavoidable when sensitive data, such as …

How does information bottleneck help deep learning?

K Kawaguchi, Z Deng, X Ji… - … Conference on Machine …, 2023 - proceedings.mlr.press
Numerous deep learning algorithms have been inspired by and understood via the notion of
information bottleneck, where unnecessary information is (often implicitly) minimized while …