Unixcoder: Unified cross-modal pre-training for code representation

D Guo, S Lu, N Duan, Y Wang, M Zhou… - arXiv preprint arXiv …, 2022 - arxiv.org
Pre-trained models for programming languages have recently demonstrated great success
on code intelligence. To support both code-related understanding and generation tasks …

Graphcodebert: Pre-training code representations with data flow

D Guo, S Ren, S Lu, Z Feng, D Tang, S Liu… - arXiv preprint arXiv …, 2020 - arxiv.org
Pre-trained models for programming language have achieved dramatic empirical
improvements on a variety of code-related tasks such as code search, code completion …

Codebert: A pre-trained model for programming and natural languages

Z Feng, D Guo, D Tang, N Duan, X Feng… - arXiv preprint arXiv …, 2020 - arxiv.org
We present CodeBERT, a bimodal pre-trained model for programming language (PL) and
nat-ural language (NL). CodeBERT learns general-purpose representations that support …

Codebleu: a method for automatic evaluation of code synthesis

S Ren, D Guo, S Lu, L Zhou, S Liu, D Tang… - arXiv preprint arXiv …, 2020 - arxiv.org
Evaluation metrics play a vital role in the growth of an area as it defines the standard of
distinguishing between good and bad models. In the area of code synthesis, the commonly …

Multi-task learning based pre-trained language model for code completion

F Liu, G Li, Y Zhao, Z Jin - Proceedings of the 35th IEEE/ACM …, 2020 - dl.acm.org
Code completion is one of the most useful features in the Integrated Development
Environments (IDEs), which can accelerate software development by suggesting the next …

Contrabert: Enhancing code pre-trained models via contrastive learning

S Liu, B Wu, X Xie, G Meng, Y Liu - 2023 IEEE/ACM 45th …, 2023 - ieeexplore.ieee.org
Large-scale pre-trained models such as CodeBERT, GraphCodeBERT have earned
widespread attention from both academia and industry. Attributed to the superior ability in …

Peculiar: Smart contract vulnerability detection based on crucial data flow graph and pre-training techniques

H Wu, Z Zhang, S Wang, Y Lei, B Lin… - 2021 IEEE 32nd …, 2021 - ieeexplore.ieee.org
Smart contracts with natural economic attributes have been widely and rapidly developed in
various fields. However, the bugs and vulnerabilities in smart contracts have brought huge …

What do pre-trained code models know about code?

A Karmakar, R Robbes - 2021 36th IEEE/ACM International …, 2021 - ieeexplore.ieee.org
Pre-trained models of code built on the transformer architecture have performed well on
software engineering (SE) tasks such as predictive code generation, code summarization …

Deep learning based code generation methods: A literature review

Z Yang, S Chen, C Gao, Z Li, G Li, R Lv - arXiv preprint arXiv:2303.01056, 2023 - arxiv.org
Code Generation aims at generating relevant code fragments according to given natural
language descriptions. In the process of software development, there exist a large number of …

Contrastive code representation learning

P Jain, A Jain, T Zhang, P Abbeel, JE Gonzalez… - arXiv preprint arXiv …, 2020 - arxiv.org
Recent work learns contextual representations of source code by reconstructing tokens from
their context. For downstream semantic understanding tasks like summarizing code in …