A review of the gumbel-max trick and its extensions for discrete stochasticity in machine learning

IAM Huijben, W Kool, MB Paulus… - IEEE transactions on …, 2022 - ieeexplore.ieee.org
The Gumbel-max trick is a method to draw a sample from a categorical distribution, given by
its unnormalized (log-) probabilities. Over the past years, the machine learning community …

Matching algorithms: Fundamentals, applications and challenges

J Ren, F Xia, X Chen, J Liu, M Hou… - … on Emerging Topics …, 2021 - ieeexplore.ieee.org
Matching plays a vital role in the rational allocation of resources in many areas, ranging from
market operation to people's daily lives. In economics, the term matching theory is coined for …

Lasuie: Unifying information extraction with latent adaptive structure-aware generative language model

H Fei, S Wu, J Li, B Li, F Li, L Qin… - Advances in …, 2022 - proceedings.neurips.cc
Universally modeling all typical information extraction tasks (UIE) with one generative
language model (GLM) has revealed great potential by the latest study, where various IE …

Discrete opinion tree induction for aspect-based sentiment analysis

C Chen, Z Teng, Z Wang, Y Zhang - … of the 60th Annual Meeting of …, 2022 - aclanthology.org
Dependency trees have been intensively used with graph neural networks for aspect-based
sentiment classification. Though being effective, such methods rely on external dependency …

Learning to compose dynamic tree structures for visual contexts

K Tang, H Zhang, B Wu, W Luo… - Proceedings of the IEEE …, 2019 - openaccess.thecvf.com
We propose to compose dynamic tree structures that place the objects in an image into a
visual context, helping visual reasoning tasks such as scene graph generation and visual …

The natural language decathlon: Multitask learning as question answering

B McCann, NS Keskar, C Xiong, R Socher - arXiv preprint arXiv …, 2018 - arxiv.org
Deep learning has improved performance on many natural language processing (NLP)
tasks individually. However, general NLP models cannot emerge within a paradigm that …

Ordered neurons: Integrating tree structures into recurrent neural networks

Y Shen, S Tan, A Sordoni, A Courville - arXiv preprint arXiv:1810.09536, 2018 - arxiv.org
Natural language is hierarchically structured: smaller units (eg, phrases) are nested within
larger units (eg, clauses). When a larger constituent ends, all of the smaller constituents that …

Star-transformer

Q Guo, X Qiu, P Liu, Y Shao, X Xue, Z Zhang - arXiv preprint arXiv …, 2019 - arxiv.org
Although Transformer has achieved great successes on many NLP tasks, its heavy structure
with fully-connected attention connections leads to dependencies on large training data. In …

Multi-pointer co-attention networks for recommendation

Y Tay, AT Luu, SC Hui - Proceedings of the 24th ACM SIGKDD …, 2018 - dl.acm.org
Many recent state-of-the-art recommender systems such as D-ATT, TransNet and
DeepCoNN exploit reviews for representation learning. This paper proposes a new neural …

Inducing target-specific latent structures for aspect sentiment classification

C Chen, Z Teng, Y Zhang - … of the 2020 conference on empirical …, 2020 - aclanthology.org
Aspect-level sentiment analysis aims to recognize the sentiment polarity of an aspect or a
target in a comment. Recently, graph convolutional networks based on linguistic …