Masked language model scoring

J Salazar, D Liang, TQ Nguyen, K Kirchhoff - arXiv preprint arXiv …, 2019 - arxiv.org
Pretrained masked language models (MLMs) require finetuning for most NLP tasks. Instead,
we evaluate MLMs out of the box via their pseudo-log-likelihood scores (PLLs), which are …

Contrastive learning with adversarial perturbations for conditional text generation

S Lee, DB Lee, SJ Hwang - arXiv preprint arXiv:2012.07280, 2020 - arxiv.org
Recently, sequence-to-sequence (seq2seq) models with the Transformer architecture have
achieved remarkable performance on various conditional text generation tasks, such as …

Weakly supervised contrastive learning for chest x-ray report generation

A Yan, Z He, X Lu, J Du, E Chang, A Gentili… - arXiv preprint arXiv …, 2021 - arxiv.org
Radiology report generation aims at generating descriptive text from radiology images
automatically, which may present an opportunity to improve radiology reporting and …

Personalized showcases: Generating multi-modal explanations for recommendations

A Yan, Z He, J Li, T Zhang, J McAuley - Proceedings of the 46th …, 2023 - dl.acm.org
Existing explanation models generate only text for recommendations but still struggle to
produce diverse contents. In this paper, to further enrich explanations, we propose a new …

Personalized complementary product recommendation

A Yan, C Dong, Y Gao, J Fu, T Zhao, Y Sun… - … Proceedings of the …, 2022 - dl.acm.org
Complementary product recommendation aims at providing product suggestions that are
often bought together to serve a joint demand. Existing work mainly focuses on modeling …

Reducing word omission errors in neural machine translation: A contrastive learning approach

Z Yang, Y Cheng, Y Liu, M Sun - … of the 57th Annual Meeting of …, 2019 - aclanthology.org
While neural machine translation (NMT) has achieved remarkable success, NMT systems
are prone to make word omission errors. In this work, we propose a contrastive learning …

Asr rescoring and confidence estimation with electra

H Futami, H Inaguma, M Mimura… - 2021 IEEE Automatic …, 2021 - ieeexplore.ieee.org
In automatic speech recognition (ASR) rescoring, the hypothesis with the fewest errors
should be selected from the n-best list using a language model (LM). However, LMs are …

Increasing visual awareness in multimodal neural machine translation from an information theoretic perspective

B Ji, T Zhang, Y Zou, B Hu, S Shen - arXiv preprint arXiv:2210.08478, 2022 - arxiv.org
Multimodal machine translation (MMT) aims to improve translation quality by equipping the
source sentence with its corresponding image. Despite the promising performance, MMT …

FILM: How can Few-Shot Image Classification Benefit from Pre-Trained Language Models?

Z Jiang, Y Dang, D Pang, H Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Few-shot learning aims to train models that can be generalized to novel classes with only a
few samples. Recently, a line of works are proposed to enhance few-shot learning with …

Learning with contrastive examples for data-to-text generation

Y Uehara, T Ishigaki, K Aoki, H Noji… - Proceedings of the …, 2020 - aclanthology.org
Existing models for data-to-text tasks generate fluent but sometimes incorrect sentences
eg,“Nikkei gains” is generated when “Nikkei drops” is expected. We investigate models …