Transformer-Patcher: One Mistake worth One Neuron Z Huang, Y Shen, X Zhang, J Zhou, W Rong, Z Xiong arXiv preprint arXiv:2301.09785, 2023 | 102 | 2023 |
Prompt-Based Meta-Learning For Few-shot Text Classification H Zhang, X Zhang, H Huang, L Yu Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022 | 28 | 2022 |
Mixture of attention heads: Selecting attention heads per token X Zhang, Y Shen, Z Huang, J Zhou, W Rong, Z Xiong arXiv preprint arXiv:2210.05144, 2022 | 24 | 2022 |
How to Determine the Most Powerful Pre-trained Language Model without Brute Force Fine-tuning? An Empirical Survey J Bai, X Zhang, C Li, H Hong, X Xu, C Lin, W Rong arXiv preprint arXiv:2312.04775, 2023 | 4 | 2023 |
Token Relation Aware Chinese Named Entity Recognition Z Huang, W Rong, X Zhang, Y Ouyang, C Lin, Z Xiong ACM Transactions on Asian and Low-Resource Language Information Processing …, 2022 | 4 | 2022 |
Solving Math Word Problems Following Logically Consistent Template Z Huang, X Zhang, J Bai, W Rong, Y Ouyang, Z Xiong 2023 International Joint Conference on Neural Networks (IJCNN), 01-08, 2023 | 1 | 2023 |