Capture, learning, and synthesis of 3D speaking styles D Cudeiro, T Bolkart, C Laidlaw, A Ranjan, MJ Black Proceedings of the IEEE/CVF conference on computer vision and pattern …, 2019 | 389 | 2019 |
Functional Adversarial Attacks C Laidlaw, S Feizi Advances in Neural Information Processing Systems 32, 2019 | 299* | 2019 |
Perceptual adversarial robustness: Defense against unseen threat models C Laidlaw, S Singla, S Feizi arXiv preprint arXiv:2006.12655, 2020 | 233 | 2020 |
Distributional preference learning: Understanding and accounting for hidden context in RLHF A Siththaranjan, C Laidlaw, D Hadfield-Menell arXiv preprint arXiv:2312.08358, 2023 | 40* | 2023 |
The boltzmann policy distribution: Accounting for systematic suboptimality in human models C Laidlaw, A Dragan arXiv preprint arXiv:2204.10759, 2022 | 34 | 2022 |
Bridging rl theory and practice with the effective horizon C Laidlaw, SJ Russell, A Dragan Advances in Neural Information Processing Systems 36, 58953-59007, 2023 | 23 | 2023 |
Playing it safe: Adversarial robustness with an abstain option C Laidlaw, S Feizi arXiv preprint arXiv:1911.11253, 2019 | 22 | 2019 |
Uncertain decisions facilitate better preference learning C Laidlaw, S Russell Advances in Neural Information Processing Systems 34, 15070-15083, 2021 | 15* | 2021 |
Preventing reward hacking with occupancy measure regularization C Laidlaw, S Singhal, A Dragan arXiv preprint arXiv:2403.03185, 2024 | 12 | 2024 |
The Effective Horizon Explains Deep RL Performance in Stochastic Environments C Laidlaw, B Zhu, S Russell, A Dragan arXiv preprint arXiv:2312.08369, 2023 | 3 | 2023 |
Toward computationally efficient inverse reinforcement learning via reward shaping LH Cooke, H Klyne, E Zhang, C Laidlaw, M Tambe, F Doshi-Velez arXiv preprint arXiv:2312.09983, 2023 | 2 | 2023 |
Scalable Oversight by Accounting for Unreliable Feedback S Singhal, C Laidlaw, A Dragan ICML 2024 Workshop on Models of Human Feedback for AI Alignment, 2024 | 1 | 2024 |
Scalably Solving Assistance Games C Laidlaw, E Bronstein, T Guo, D Feng, L Berglund, J Svegliato, S Russell, ... ICML 2024 Workshop on Models of Human Feedback for AI Alignment, 2024 | 1 | 2024 |