Natural language generation for effective knowledge distillation

R Tang, Y Lu, J Lin - Proceedings of the 2nd Workshop on Deep …, 2019 - aclanthology.org
Abstract Knowledge distillation can effectively transfer knowledge from BERT, a deep
language representation model, to traditional, shallow word embedding-based neural …

[PDF][PDF] Natural Language Generation for Effective Knowledge Distillation

R Tang, Y Lu, J Lin - cs.uwaterloo.ca
Abstract Knowledge distillation can effectively transfer knowledge from BERT, a deep
language representation model, to traditional, shallow word embedding-based neural …

[PDF][PDF] Natural Language Generation for Effective Knowledge Distillation

R Tang, Y Lu, J Lin - ralphtang.com
Abstract Knowledge distillation can effectively transfer knowledge from BERT, a deep
language representation model, to traditional, shallow word embedding-based neural …

[PDF][PDF] Natural Language Generation for Effective Knowledge Distillation

R Tang, Y Lu, J Lin - EMNLP-IJCNLP 2019, 2019 - aclanthology.org
Abstract Knowledge distillation can effectively transfer knowledge from BERT, a deep
language representation model, to traditional, shallow word embedding-based neural …