Improved code summarization via a graph neural network

A LeClair, S Haque, L Wu, C McMillan - Proceedings of the 28th …, 2020 - dl.acm.org
Automatic source code summarization is the task of generating natural language
descriptions for source code. Automatic code summarization is a rapidly expanding research …

Universal dependency parsing from scratch

P Qi, T Dozat, Y Zhang, CD Manning - arXiv preprint arXiv:1901.10457, 2019 - arxiv.org
This paper describes Stanford's system at the CoNLL 2018 UD Shared Task. We introduce a
complete neural pipeline system that takes raw text as input, and performs all tasks required …

Glyce: Glyph-vectors for chinese character representations

Y Meng, W Wu, F Wang, X Li, P Nie… - Advances in …, 2019 - proceedings.neurips.cc
It is intuitive that NLP tasks for logographic languages like Chinese should benefit from the
use of the glyph information in those languages. However, due to the lack of rich …

Semantic neural machine translation using AMR

L Song, D Gildea, Y Zhang, Z Wang… - Transactions of the …, 2019 - direct.mit.edu
It is intuitive that semantic representations can be useful for machine translation, mainly
because they can help in enforcing meaning preservation and handling data sparsity (many …

Transforming question answering datasets into natural language inference datasets

D Demszky, K Guu, P Liang - arXiv preprint arXiv:1809.02922, 2018 - arxiv.org
Existing datasets for natural language inference (NLI) have propelled research on language
understanding. We propose a new method for automatically deriving NLI datasets from the …

Is word segmentation necessary for deep learning of Chinese representations?

X Li, Y Meng, X Sun, Q Han, A Yuan, J Li - arXiv preprint arXiv:1905.05526, 2019 - arxiv.org
Segmenting a chunk of text into words is usually the first step of processing Chinese text, but
its necessity has rarely been explored. In this paper, we ask the fundamental question of …

Generating sentences from disentangled syntactic and semantic spaces

Y Bao, H Zhou, S Huang, L Li, L Mou… - arXiv preprint arXiv …, 2019 - arxiv.org
Variational auto-encoders (VAEs) are widely used in natural language generation due to the
regularization of the latent space. However, generating sentences from the continuous latent …

Does syntax need to grow on trees? sources of hierarchical inductive bias in sequence-to-sequence networks

RT McCoy, R Frank, T Linzen - Transactions of the Association for …, 2020 - direct.mit.edu
Learners that are exposed to the same training data might generalize differently due to
differing inductive biases. In neural network models, inductive biases could in theory arise …

[PDF][PDF] 神经机器翻译综述

李亚超, 熊德意, 张民 - 计算机学报, 2018 - cdn.jsdelivr.net
摘要机器翻译研究将源语言所表达的语义自动转换为目标语言的相同语义,
是人工智能和自然语言处理的重要研究内容. 近年来, 基于序列到序列模型(Sequence-to …

Hierarchy-aware global model for hierarchical text classification

J Zhou, C Ma, D Long, G Xu, N Ding… - Proceedings of the …, 2020 - aclanthology.org
Hierarchical text classification is an essential yet challenging subtask of multi-label text
classification with a taxonomic hierarchy. Existing methods have difficulties in modeling the …