LRC-BERT: latent-representation contrastive knowledge distillation for natural language understanding

H Fu, S Zhou, Q Yang, J Tang, G Liu, K Liu… - Proceedings of the AAAI …, 2021 - ojs.aaai.org
The pre-training models such as BERT have achieved great results in various natural
language processing problems. However, a large number of parameters need significant …

[PDF][PDF] LRC-BERT: Latent-representation Contrastive Knowledge Distillation for Natural Language Understanding

H Fu, S Zhou, Q Yang, J Tang, G Liu, K Liu, X Li - 2021 - bigdata.ustc.edu.cn
The pre-training models such as BERT have achieved great results in various natural
language processing problems. However, a large number of parameters need significant …

[PDF][PDF] LRC-BERT: Latent-representation Contrastive Knowledge Distillation for Natural Language Understanding

H Fu, S Zhou, Q Yang, J Tang, G Liu, K Liu, X Li - 2021 - cdn.aaai.org
The pre-training models such as BERT have achieved great results in various natural
language processing problems. However, a large number of parameters need significant …

LRC-BERT: Latent-representation Contrastive Knowledge Distillation for Natural Language Understanding

H Fu, S Zhou, Q Yang, J Tang, G Liu, K Liu… - arXiv e …, 2020 - ui.adsabs.harvard.edu
The pre-training models such as BERT have achieved great results in various natural
language processing problems. However, a large number of parameters need significant …

LRC-BERT: Latent-representation Contrastive Knowledge Distillation for Natural Language Understanding

H Fu, S Zhou, Q Yang, J Tang, G Liu, K Liu… - arXiv preprint arXiv …, 2020 - arxiv.org
The pre-training models such as BERT have achieved great results in various natural
language processing problems. However, a large number of parameters need significant …

[PDF][PDF] LRC-BERT: Latent-representation Contrastive Knowledge Distillation for Natural Language Understanding

H Fu, S Zhou, Q Yang, J Tang, G Liu, K Liu, X Li - 2021 - scholar.archive.org
The pre-training models such as BERT have achieved great results in various natural
language processing problems. However, a large number of parameters need significant …

[PDF][PDF] LRC-BERT: Latent-representation Contrastive Knowledge Distillation for Natural Language Understanding

H Fu, S Zhou, Q Yang, J Tang, G Liu, K Liu, X Li - 2021 - dm.ustc.edu.cn
The pre-training models such as BERT have achieved great results in various natural
language processing problems. However, a large number of parameters need significant …