Fairly Evaluating Large Language Model-based Recommendation Needs Revisit the Cross-Entropy Loss

C Xu, Z Zhu, J Wang, J Wang, W Zhang - arXiv preprint arXiv:2402.06216, 2024 - arxiv.org
Large language models (LLMs) have gained much attention in the recommendation
community; some studies have observed that LLMs, fine-tuned by the cross-entropy loss …

Self-normalized importance sampling for neural language modeling

Z Yang, Y Gao, A Gerstenberger, J Jiang… - arXiv preprint arXiv …, 2021 - arxiv.org
To mitigate the problem of having to traverse over the full vocabulary in the softmax
normalization of a neural language model, sampling-based training criteria are proposed …

[PDF][PDF] Language modeling and machine translation: improvements in training and modeling

G Yingbo - www-i6.informatik.rwth-aachen.de
The field of statistical language modeling and machine translation has seen rapid
developments in recent years, with artificial neural networks taking center of the stage …