Graph neural networks: A review of methods and applications J Zhou, G Cui, S Hu, Z Zhang, C Yang, Z Liu, L Wang, C Li, M Sun AI open 1, 57-81, 2020 | 5579 | 2020 |
Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification S Hu, N Ding, H Wang, Z Liu, J Li, M Sun | 299 | 2021 |
Parameter-efficient fine-tuning of large-scale pre-trained language models N Ding, Y Qin, G Yang, F Wei, Z Yang, Y Su, S Hu, Y Chen, CM Chan, ... Nature Machine Intelligence 5 (3), 220-235, 2023 | 273 | 2023 |
Openprompt: An open-source framework for prompt-learning N Ding, S Hu, W Zhao, Y Chen, Z Liu, HT Zheng, M Sun arXiv preprint arXiv:2111.01998, 2021 | 229 | 2021 |
Delta tuning: A comprehensive study of parameter efficient methods for pre-trained language models N Ding, Y Qin, G Yang, F Wei, Z Yang, Y Su, S Hu, Y Chen, CM Chan, ... arXiv preprint arXiv:2203.06904, 2022 | 181 | 2022 |
Tool learning with foundation models Y Qin, S Hu, Y Lin, W Chen, N Ding, G Cui, Z Zeng, Y Huang, C Xiao, ... arXiv preprint arXiv:2304.08354, 2023 | 175 | 2023 |
Enhancing chat language models by scaling high-quality instructional conversations N Ding, Y Chen, B Xu, Y Qin, Z Zheng, S Hu, Z Liu, M Sun, B Zhou arXiv preprint arXiv:2305.14233, 2023 | 141 | 2023 |
Prototypical verbalizer for prompt-based few-shot tuning G Cui, S Hu, N Ding, L Huang, Z Liu arXiv preprint arXiv:2203.09770, 2022 | 80 | 2022 |
Graph Policy Network for Transferable Active Learning on Graphs S Hu, Z Xiong, M Qu, X Yuan, MA Côté, Z Liu, J Tang NeurIPS'20, 2020 | 58 | 2020 |
Copen: Probing conceptual knowledge in pre-trained language models H Peng, X Wang, S Hu, H Jin, L Hou, J Li, Z Liu, Q Liu arXiv preprint arXiv:2211.04079, 2022 | 24 | 2022 |
Decoder-only or encoder-decoder? interpreting language model as a regularized encoder-decoder Z Fu, W Lam, Q Yu, AMC So, S Hu, Z Liu, N Collier arXiv preprint arXiv:2304.04052, 2023 | 17 | 2023 |
Won't Get Fooled Again: Answering Questions with False Premises S Hu, Y Luo, H Wang, X Cheng, Z Liu, M Sun arXiv preprint arXiv:2307.02394, 2023 | 9 | 2023 |
Sparse structure search for delta tuning S Hu, Z Zhang, N Ding, Y Wang, Y Wang, Z Liu, M Sun Advances in Neural Information Processing Systems 35, 9853-9865, 2022 | 9 | 2022 |
OpenDelta: A Plug-and-play Library for Parameter-efficient Adaptation of Pre-trained Models S Hu, N Ding, W Zhao, X Lv, Z Zhang, Z Liu, M Sun arXiv preprint arXiv:2307.03084, 2023 | 8 | 2023 |
Sparse structure search for parameter-efficient tuning S Hu, Z Zhang, N Ding, Y Wang, Y Wang, Z Liu, M Sun arXiv preprint arXiv:2206.07382, 2022 | 8 | 2022 |
Minicpm: Unveiling the potential of small language models with scalable training strategies S Hu, Y Tu, X Han, C He, G Cui, X Long, Z Zheng, Y Fang, Y Huang, ... arXiv preprint arXiv:2404.06395, 2024 | 7 | 2024 |
OlympiadBench: A Challenging Benchmark for Promoting AGI with Olympiad-Level Bilingual Multimodal Scientific Problems C He, R Luo, Y Bai, S Hu, ZL Thai, J Shen, J Hu, X Han, Y Huang, ... arXiv preprint arXiv:2402.14008, 2024 | 5 | 2024 |
KACC: A multi-task benchmark for knowledge abstraction, concretization and completion J Zhou, S Hu, X Lv, C Yang, Z Liu, W Xu, J Jiang, J Li, M Sun arXiv preprint arXiv:2004.13631, 2020 | 4 | 2020 |
Exploring Lottery Prompts for Pre-trained Language Models Y Chen, N Ding, X Wang, S Hu, HT Zheng, Z Liu, P Xie arXiv preprint arXiv:2305.19500, 2023 | 3 | 2023 |
Unified View of Grokking, Double Descent and Emergent Abilities: A Perspective from Circuits Competition Y Huang, S Hu, X Han, Z Liu, M Sun arXiv preprint arXiv:2402.15175, 2024 | 2 | 2024 |