Multimodal machine learning is a vibrant multi-disciplinary research field that aims to design computer agents with intelligent capabilities such as understanding, reasoning, and learning …
Prompting has shown impressive success in enabling large pretrained language models (LMs) to perform diverse NLP tasks, especially when only few downstream data are …
Multi-dimensional evaluation is the dominant paradigm for human evaluation in Natural Language Generation (NLG), ie, evaluating the generated text from multiple explainable …
S Gehrmann, E Clark, T Sellam - Journal of Artificial Intelligence Research, 2023 - jair.org
Abstract Evaluation practices in natural language generation (NLG) have many known flaws, but improved evaluation approaches are rarely widely adopted. This issue has become …
The ability of large language models (LLMs) to follow natural language instructions with human-level fluency suggests many opportunities in healthcare to reduce administrative …
Y Zha, Y Yang, R Li, Z Hu - arXiv preprint arXiv:2305.16739, 2023 - arxiv.org
Many text generation applications require the generated text to be factually consistent with input information. Automatic evaluation of factual consistency is challenging. Previous work …
H Ye, N Zhang, H Chen, H Chen - arXiv preprint arXiv:2210.12714, 2022 - arxiv.org
Generative Knowledge Graph Construction (KGC) refers to those methods that leverage the sequence-to-sequence framework for building knowledge graphs, which is flexible and can …
Y Zha, Y Yang, R Li, Z Hu - Advances in Neural Information …, 2023 - proceedings.neurips.cc
Large language models (LLMs), typically designed as a function of next-word prediction, have excelled across extensive NLP tasks. Despite the generality, next-word prediction is …
Multi-step reasoning ability is fundamental to many natural language tasks, yet it is unclear what constitutes a good reasoning chain and how to evaluate them. Most existing methods …