Recipes for building an open-domain chatbot S Roller, E Dinan, N Goyal, D Ju, M Williamson, Y Liu, J Xu, M Ott, ... arXiv preprint arXiv:2004.13637, 2020 | 995 | 2020 |
Blenderbot 3: a deployed conversational agent that continually learns to responsibly engage K Shuster, J Xu, M Komeili, D Ju, EM Smith, S Roller, M Ung, M Chen, ... arXiv preprint arXiv:2208.03188, 2022 | 233 | 2022 |
Beyond goldfish memory: Long-term open-domain conversation J Xu, A Szlam, J Weston arXiv preprint arXiv:2107.07567, 2021 | 218 | 2021 |
Recipes for safety in open-domain chatbots J Xu, D Ju, M Li, YL Boureau, J Weston, E Dinan arXiv preprint arXiv:2010.07079, 2020 | 167 | 2020 |
Chain-of-verification reduces hallucination in large language models S Dhuliawala, M Komeili, J Xu, R Raileanu, X Li, A Celikyilmaz, J Weston arXiv preprint arXiv:2309.11495, 2023 | 142 | 2023 |
Self-rewarding language models W Yuan, RY Pang, K Cho, S Sukhbaatar, J Xu, J Weston arXiv preprint arXiv:2401.10020, 2024 | 120 | 2024 |
Bot-adversarial dialogue for safe conversational agents J Xu, D Ju, M Li, YL Boureau, J Weston, E Dinan Proceedings of the 2021 Conference of the North American Chapter of the …, 2021 | 110 | 2021 |
Learning new skills after deployment: Improving open-domain internet-driven dialogue with human feedback J Xu, M Ung, M Komeili, K Arora, YL Boureau, J Weston arXiv preprint arXiv:2208.03270, 2022 | 33 | 2022 |
Saferdialogues: Taking feedback gracefully after conversational safety failures M Ung, J Xu, YL Boureau arXiv preprint arXiv:2110.07518, 2021 | 31 | 2021 |
Some things are more cringe than others: Preference optimization with the pairwise cringe loss J Xu, A Lee, S Sukhbaatar, J Weston arXiv preprint arXiv:2312.16682, 2023 | 25 | 2023 |
The cringe loss: Learning what language not to model L Adolphs, T Gao, J Xu, K Shuster, S Sukhbaatar, J Weston arXiv preprint arXiv:2211.05826, 2022 | 25 | 2022 |
On anytime learning at macroscale L Caccia, J Xu, M Ott, M Ranzato, L Denoyer Conference on Lifelong Learning Agents, 165-182, 2022 | 22 | 2022 |
When life gives you lemons, make cherryade: Converting feedback from bad responses into good labels W Shi, E Dinan, K Shuster, J Weston, J Xu arXiv preprint arXiv:2210.15893, 2022 | 15 | 2022 |
Learning from data in the mixed adversarial non-adversarial case: Finding the helpers and ignoring the trolls D Ju, J Xu, YL Boureau, J Weston arXiv preprint arXiv:2208.03295, 2022 | 13 | 2022 |
Housing choices, sorting, and the distribution of educational benefits under deferred acceptance J Xu Journal of Public Economic Theory 21 (3), 558-595, 2019 | 8 | 2019 |
Training models to generate, recognize, and reframe unhelpful thoughts M Maddela, M Ung, J Xu, A Madotto, H Foran, YL Boureau arXiv preprint arXiv:2307.02768, 2023 | 6 | 2023 |
Improving open language models by learning from organic interactions J Xu, D Ju, J Lane, M Komeili, EM Smith, M Ung, M Behrooz, W Ngan, ... arXiv preprint arXiv:2306.04707, 2023 | 5 | 2023 |
Parameter estimation in gaussian mixture models with malicious noise, without balanced mixing coefficients J Xu, J Mareček 2018 56th Annual Allerton Conference on Communication, Control, and …, 2018 | 4 | 2018 |
Following Length Constraints in Instructions W Yuan, I Kulikov, P Yu, K Cho, S Sukhbaatar, J Weston, J Xu arXiv preprint arXiv:2406.17744, 2024 | | 2024 |
Continual Dialogue State Tracking via Example-Guided Question Answering H Cho, A Madotto, Z Lin, KR Chandu, S Kottur, J Xu, J May, C Sankar arXiv preprint arXiv:2305.13721, 2023 | | 2023 |