In-context examples selection for machine translation

S Agrawal, C Zhou, M Lewis, L Zettlemoyer… - arXiv preprint arXiv …, 2022 - arxiv.org
Large-scale generative models show an impressive ability to perform a wide range of
Natural Language Processing (NLP) tasks using in-context learning, where a few examples …

Repocoder: Repository-level code completion through iterative retrieval and generation

F Zhang, B Chen, Y Zhang, J Keung, J Liu… - arXiv preprint arXiv …, 2023 - arxiv.org
The task of repository-level code completion is to continue writing the unfinished code based
on a broader context of the repository. While for automated code completion tools, it is …

Lift yourself up: Retrieval-augmented text generation with self-memory

X Cheng, D Luo, X Chen, L Liu… - Advances in Neural …, 2024 - proceedings.neurips.cc
With direct access to human-written reference as memory, retrieval-augmented generation
has achieved much progress in a wide range of text generation tasks. Since better memory …

[PDF][PDF] Improving few-shot prompts with relevant static analysis products

T Ahmed, KS Pai, P Devanbu, ET Barr - arXiv preprint arXiv …, 2023 - academia.edu
ABSTRACT Large Language Models (LLM) are a new class of computation engines,
łprogrammedž via prompt engineering. Researchers are still learning how to best łprogramž …

Automatic semantic augmentation of language model prompts (for code summarization)

T Ahmed, KS Pai, P Devanbu, E Barr - Proceedings of the IEEE/ACM …, 2024 - dl.acm.org
Large Language Models (LLM) are a new class of computation engines," programmed" via
prompt engineering. Researchers are still learning how to best" program" these LLMs to …

CTQScorer: Combining multiple features for in-context example selection for machine translation

A Kumar, R Puduppully, R Dabre… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models have demonstrated the capability to perform on machine translation
when the input is prompted with a few examples (in-context learning). Translation quality …

The pipeline system of asr and nlu with mlm-based data augmentation toward stop low-resource challenge

H Futami, J Huynh, S Arora, SL Wu… - ICASSP 2023-2023 …, 2023 - ieeexplore.ieee.org
This paper describes our system for the low-resource domain adaptation track (Track 3) in
Spoken Language Understanding Grand Challenge, which is a part of ICASSP Signal …

A Statistical Framework for Data-dependent Retrieval-Augmented Models

S Basu, AS Rawat, M Zaheer - arXiv preprint arXiv:2408.15399, 2024 - arxiv.org
Modern ML systems increasingly augment input instances with additional relevant
information to enhance final prediction. Despite growing interest in such retrieval …

Automatic Combination of Sample Selection Strategies for Few-Shot Learning

B Pecher, I Srba, M Bielikova, J Vanschoren - arXiv preprint arXiv …, 2024 - arxiv.org
In few-shot learning, such as meta-learning, few-shot fine-tuning or in-context learning, the
limited number of samples used to train a model have a significant impact on the overall …

RepoMinCoder: Improving Repository-Level Code Generation Based on Information Loss Screening

Y Li, E Shi, D Zheng, K Duan, J Chen… - Proceedings of the 15th …, 2024 - dl.acm.org
Repository-level code generation task involves generating code at a specified location
based on unfinished code with repository context. Existing research mainly rely on retrieval …