Bridging the gap: Attending to discontinuity in identification of multiword expressions

O Rohanian, S Taslimipoor, S Kouchaki, LA Ha… - arXiv preprint arXiv …, 2019 - arxiv.org
arXiv preprint arXiv:1902.10667, 2019arxiv.org
We introduce a new method to tag Multiword Expressions (MWEs) using a linguistically
interpretable language-independent deep learning architecture. We specifically target
discontinuity, an under-explored aspect that poses a significant challenge to computational
treatment of MWEs. Two neural architectures are explored: Graph Convolutional Network
(GCN) and multi-head self-attention. GCN leverages dependency parse information, and
self-attention attends to long-range relations. We finally propose a combined model that …
We introduce a new method to tag Multiword Expressions (MWEs) using a linguistically interpretable language-independent deep learning architecture. We specifically target discontinuity, an under-explored aspect that poses a significant challenge to computational treatment of MWEs. Two neural architectures are explored: Graph Convolutional Network (GCN) and multi-head self-attention. GCN leverages dependency parse information, and self-attention attends to long-range relations. We finally propose a combined model that integrates complementary information from both through a gating mechanism. The experiments on a standard multilingual dataset for verbal MWEs show that our model outperforms the baselines not only in the case of discontinuous MWEs but also in overall F-score.
arxiv.org
以上显示的是最相近的搜索结果。 查看全部搜索结果