Obserwuj
Yiming Cui
Yiming Cui
Research Scientist, iFLYTEK Research
Zweryfikowany adres z iflytek.com - Strona główna
Tytuł
Cytowane przez
Cytowane przez
Rok
Pre-training With Whole Word Masking for Chinese BERT
Y Cui, W Che, T Liu, B Qin, Z Yang
IEEE/ACM Transactions on Audio, Speech, and Language Processing (TASLP) 29 …, 2021
12072021
Revisiting Pre-Trained Models for Chinese Natural Language Processing
Y Cui, W Che, T Liu, B Qin, S Wang, G Hu
Findings of EMNLP 2020, 657–668, 2020
6372020
Attention-over-Attention Neural Networks for Reading Comprehension
Y Cui, Z Chen, S Wei, S Wang, T Liu, G Hu
ACL 2017, 593–602, 2017
5272017
CLUE: A Chinese Language Understanding Evaluation Benchmark
L Xu, X Zhang, L Li, H Hu, C Cao, W Liu, J Li, Y Li, K Sun, Y Xu, Y Cui, ...
COLING 2020, 4762–4772, 2020
2892020
A Span-Extraction Dataset for Chinese Machine Reading Comprehension
Y Cui, T Liu, L Xiao, Z Chen, W Ma, W Che, S Wang, G Hu
EMNLP-IJCNLP 2019, 593–602, 2018
1942018
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting
S Chen, Y Hou, Y Cui, W Che, T Liu, X Yu
EMNLP 2020, 7870-–7881, 2020
1452020
Exploiting Persona Information for Diverse Generation of Conversational Responses
H Song, WN Zhang, Y Cui, D Wang, T Liu
IJCAI-19, 5190–5196, 2019
1272019
Consensus Attention-based Neural Networks for Chinese Reading Comprehension
Y Cui, T Liu, Z Chen, S Wang, G Hu
COLING 2016, 1777–1786, 2016
1102016
Efficient and effective text encoding for chinese llama and alpaca
Y Cui, Z Yang, X Yao
arXiv preprint arXiv:2304.08177, 2023
1092023
CharBERT: Character-aware Pre-trained Language Model
W Ma, Y Cui, C Si, T Liu, S Wang, G Hu
COLING 2020, 39–50, 2020
892020
CJRC: A Reliable Human-Annotated Benchmark DataSet for Chinese Judicial Reading Comprehension
X Duan, B Wang, Z Wang, W Ma, Y Cui, D Wu, S Wang, T Liu, T Huo, Z Hu, ...
CCL 2019, 439–451, 2019
662019
Is Graph Structure Necessary for Multi-hop Question Answering?
N Shao, Y Cui, T Liu, S Wang, G Hu
EMNLP 2020, 7187-7192, 2020
652020
Generating and Exploiting Large-scale Pseudo Training Data for Zero Pronoun Resolution
T Liu, Y Cui, Q Yin, S Wang, W Zhang, G Hu
ACL 2017, 102–111, 2017
582017
Cross-Lingual Machine Reading Comprehension
Y Cui, W Che, T Liu, B Qin, S Wang, G Hu
EMNLP-IJCNLP 2019, 1586–1595, 2019
572019
Context-Sensitive Generation of Open-Domain Conversational Responses
W Zhang, Y Cui, Y Wang, Q Zhu, L Li, L Zhou, T Liu
COLING 2018, 2437–2447, 2018
522018
LSTM Neural Reordering Feature for Statistical Machine Translation
Y Cui, S Wang, J Li
NAACL 2016, 977–982, 2016
442016
PERT: pre-training BERT with permuted language model
Y Cui, Z Yang, T Liu
arXiv preprint arXiv:2203.06906, 2022
412022
TextBrewer: An Open-Source Knowledge Distillation Toolkit for Natural Language Processing
Z Yang, Y Cui, Z Chen, W Che, T Liu, S Wang, G Hu
ACL 2020: System Demonstrations, 9–16, 2020
382020
Convolutional Spatial Attention Model for Reading Comprehension with Multiple-Choice Questions
Z Chen, Y Cui, W Ma, S Wang, G Hu
AAAI-19, 6276–6283, 2018
352018
Benchmarking Robustness of Machine Reading Comprehension Models
C Si, Z Yang, Y Cui, W Ma, T Liu, S Wang
Findings of ACL-IJCNLP 2021, 634–644, 2020
302020
Nie można teraz wykonać tej operacji. Spróbuj ponownie później.
Prace 1–20