SafeDecoding: Defending against Jailbreak Attacks via Safety-Aware Decoding Z Xu, F Jiang, L Niu, J Jia, BY Lin, R Poovendran ACL 2024, 2024 | 25 | 2024 |
ArtPrompt: ASCII Art-based Jailbreak Attacks against Aligned LLMs F Jiang, Z Xu, L Niu, Z Xiang, B Ramasubramanian, B Li, R Poovendran ACL 2024, 2024 | 16 | 2024 |
Wireless Distributed Consensus in Vehicle to Vehicle Networks for Autonomous Driving C Feng, Z Xu, X Zhu, PV Klaine, L Zhang IEEE Transactions on Vehicular Technology, 2023 | 14 | 2023 |
Identifying and Mitigating Vulnerabilities in LLM-Integrated Applications F Jiang, Z Xu, L Niu, B Wang, J Jia, B Li, R Poovendran Instruction Workshop @ NeurIPS 2023, 2023 | 7 | 2023 |
Exact Fault-Tolerant Consensus With Voting Validity Z Xu, Y Li, C Feng, L Zhang 2023 IEEE International Parallel and Distributed Processing Symposium, 2023 | 2 | 2023 |
CleanGen: Mitigating Backdoor Attacks for Generation Tasks in Large Language Models Y Li, Z Xu, F Jiang, L Niu, D Sahabandu, B Ramasubramanian, ... arXiv preprint arXiv:2406.12257, 2024 | 1 | 2024 |
ChatBug: A Common Vulnerability of Aligned LLMs Induced by Chat Templates F Jiang, Z Xu, L Niu, BY Lin, R Poovendran arXiv preprint arXiv:2406.12935, 2024 | | 2024 |
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing Z Xu, F Jiang, L Niu, Y Deng, R Poovendran, Y Choi, BY Lin arXiv preprint arXiv:2406.08464, 2024 | | 2024 |
ACE: A Model Poisoning Attack on Contribution Evaluation Methods in Federated Learning Z Xu, F Jiang, L Niu, J Jia, B Li, R Poovendran Usenix Security 2024, 2024 | | 2024 |
Brave: Byzantine-Resilient and Privacy-Preserving Peer-to-Peer Federated Learning Z Xu, F Jiang, L Niu, J Jia, R Poovendran The 5th AAAI Workshop on Privacy-Preserving Artificial Intelligence, 2024 | | 2024 |