Recent years have seen the successful application of large pre-trained models to code representation learning, resulting in substantial improvements on many code-related downstream tasks. But there are issues surrounding their application to SE tasks. First, the majority of the pre-trained models focus on pre-training only the encoder of the Transformer. For generation tasks that are addressed using models with the encoder-decoder architecture, however, there is no reason why the decoder should be left out during pre-training. Second, many existing pre-trained models, including state-of-the-art models such as T5-learning, simply reuse the pre-training tasks designed for natural languages. Moreover, to learn the natural language description ...
We provide the datasets for reproducing the experiments in "Code Execution with Pre-trained Language...
Datasets for the paper "Bridging Pre-trained Models and Downstream Tasks for Source Code Understandi...
Few-shot learning with large-scale, pre-trained language models is a powerful way to answer question...
An effective and efficient code representation is critical to the success of sequence-to-sequence de...
Back-translation is widely known for its effectiveness for neural machine translation when little to...
Representation learning of source code is essential for applying machine learning to software engine...
Code generation is a longstanding challenge, aiming to generate a code snippet based on a natural la...
Recently, many pre-trained language models for source code have been proposed to model the context o...
Machine-learning models can reach very high performance with supervised training, where they learn f...
Structure prediction (SP) tasks are important in natural language understanding in the sense that th...
We study the problem of building generative models of natural source code (NSC); that is, source cod...
There has been a recent surge of interest in automating software engineering tasks using deep learni...
Training a deep learning model on source code has gained significant traction recently. Since such m...
Learning code representations has found many uses in software engineering, such as code classificati...
An effective and efficient encoding of the source code of a computer program is critical to the succ...
We provide the datasets for reproducing the experiments in "Code Execution with Pre-trained Language...
Datasets for the paper "Bridging Pre-trained Models and Downstream Tasks for Source Code Understandi...
Few-shot learning with large-scale, pre-trained language models is a powerful way to answer question...
An effective and efficient code representation is critical to the success of sequence-to-sequence de...
Back-translation is widely known for its effectiveness for neural machine translation when little to...
Representation learning of source code is essential for applying machine learning to software engine...
Code generation is a longstanding challenge, aiming to generate a code snippet based on a natural la...
Recently, many pre-trained language models for source code have been proposed to model the context o...
Machine-learning models can reach very high performance with supervised training, where they learn f...
Structure prediction (SP) tasks are important in natural language understanding in the sense that th...
We study the problem of building generative models of natural source code (NSC); that is, source cod...
There has been a recent surge of interest in automating software engineering tasks using deep learni...
Training a deep learning model on source code has gained significant traction recently. Since such m...
Learning code representations has found many uses in software engineering, such as code classificati...
An effective and efficient encoding of the source code of a computer program is critical to the succ...
We provide the datasets for reproducing the experiments in "Code Execution with Pre-trained Language...
Datasets for the paper "Bridging Pre-trained Models and Downstream Tasks for Source Code Understandi...
Few-shot learning with large-scale, pre-trained language models is a powerful way to answer question...