关注
Yiming Cui
Yiming Cui
Research Scientist, iFLYTEK Research
在 iflytek.com 的电子邮件经过验证 - 首页
标题
引用次数
引用次数
年份
Pre-training With Whole Word Masking for Chinese BERT
Y Cui, W Che, T Liu, B Qin, Z Yang
IEEE/ACM Transactions on Audio, Speech, and Language Processing (TASLP) 29 …, 2021
13412021
Revisiting Pre-Trained Models for Chinese Natural Language Processing
Y Cui, W Che, T Liu, B Qin, S Wang, G Hu
Findings of EMNLP 2020, 657–668, 2020
7132020
Attention-over-Attention Neural Networks for Reading Comprehension
Y Cui, Z Chen, S Wei, S Wang, T Liu, G Hu
ACL 2017, 593–602, 2017
5292017
CLUE: A Chinese Language Understanding Evaluation Benchmark
L Xu, X Zhang, L Li, H Hu, C Cao, W Liu, J Li, Y Li, K Sun, Y Xu, Y Cui, ...
COLING 2020, 4762–4772, 2020
3172020
A Span-Extraction Dataset for Chinese Machine Reading Comprehension
Y Cui, T Liu, L Xiao, Z Chen, W Ma, W Che, S Wang, G Hu
EMNLP-IJCNLP 2019, 593–602, 2018
2042018
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting
S Chen, Y Hou, Y Cui, W Che, T Liu, X Yu
EMNLP 2020, 7870-–7881, 2020
1682020
Efficient and effective text encoding for chinese llama and alpaca
Y Cui, Z Yang, X Yao
arXiv preprint arXiv:2304.08177, 2023
1592023
Exploiting Persona Information for Diverse Generation of Conversational Responses
H Song, WN Zhang, Y Cui, D Wang, T Liu
IJCAI-19, 5190–5196, 2019
1312019
Consensus Attention-based Neural Networks for Chinese Reading Comprehension
Y Cui, T Liu, Z Chen, S Wang, G Hu
COLING 2016, 1777–1786, 2016
1102016
CharBERT: Character-aware Pre-trained Language Model
W Ma, Y Cui, C Si, T Liu, S Wang, G Hu
COLING 2020, 39–50, 2020
922020
CJRC: A Reliable Human-Annotated Benchmark DataSet for Chinese Judicial Reading Comprehension
X Duan, B Wang, Z Wang, W Ma, Y Cui, D Wu, S Wang, T Liu, T Huo, Z Hu, ...
CCL 2019, 439–451, 2019
742019
Is Graph Structure Necessary for Multi-hop Question Answering?
N Shao, Y Cui, T Liu, S Wang, G Hu
EMNLP 2020, 7187-7192, 2020
642020
Cross-Lingual Machine Reading Comprehension
Y Cui, W Che, T Liu, B Qin, S Wang, G Hu
EMNLP-IJCNLP 2019, 1586–1595, 2019
592019
Generating and Exploiting Large-scale Pseudo Training Data for Zero Pronoun Resolution
T Liu, Y Cui, Q Yin, S Wang, W Zhang, G Hu
ACL 2017, 102–111, 2017
572017
Context-Sensitive Generation of Open-Domain Conversational Responses
W Zhang, Y Cui, Y Wang, Q Zhu, L Li, L Zhou, T Liu
COLING 2018, 2437–2447, 2018
532018
LSTM Neural Reordering Feature for Statistical Machine Translation
Y Cui, S Wang, J Li
NAACL 2016, 977–982, 2016
482016
PERT: pre-training BERT with permuted language model
Y Cui, Z Yang, T Liu
arXiv preprint arXiv:2203.06906, 2022
442022
TextBrewer: An Open-Source Knowledge Distillation Toolkit for Natural Language Processing
Z Yang, Y Cui, Z Chen, W Che, T Liu, S Wang, G Hu
ACL 2020: System Demonstrations, 9–16, 2020
402020
Convolutional Spatial Attention Model for Reading Comprehension with Multiple-Choice Questions
Z Chen, Y Cui, W Ma, S Wang, G Hu
AAAI-19, 6276–6283, 2018
362018
CINO: A Chinese Minority Pre-trained Language Model
Z Yang, Z Xu, Y Cui, B Wang, M Lin, D Wu, Z Chen
COLING 2022, 3937–3949, 2022
302022
系统目前无法执行此操作,请稍后再试。
文章 1–20