In this work, we evaluate 10 open-source instructed LLMs on four representative code comprehension and generation tasks. We have the following main findings. First, for the zero-shot setting, instructed LLMs are very competitive on code comprehension and generation tasks and sometimes even better than small SOTA models specifically fine-tuned on each downstream task. We also find that larger instructed LLMs are not always better on code-related tasks. Second, for the few-shot setting, we find that adding demonstration examples substantially helps instructed LLMs perform better on most code comprehension and generation tasks; however, the examples would sometimes induce unstable or even worse performance. Furthermore, we find widely-used BM2...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Large language models (LLMs) have demonstrated significant potential in the realm of natural languag...
The automation of code review activities, a long-standing pursuit in software engineering, has been ...
Large language models~(LLMs) are instruction followers, but it can be challenging to find the best i...
Recently, Instruction fine-tuning has risen to prominence as a potential method for enhancing the ze...
Very large language models (LLMs), such as GPT-3 and Codex have achieved state-of-the-art performanc...
Recent work has shown that fine-tuning large pre-trained language models on a collection of tasks de...
Sparse Mixture-of-Experts (MoE) is a neural architecture design that can be utilized to add learnabl...
This paper systematically investigates the generation of code explanations by Large Language Models ...
Large Language Models (LLM) are a new class of computation engines, "programmed" via prompt engineer...
Large Language Models (LLMs) for code are a family of high-parameter, transformer-based neural netwo...
Multi-task learning (MTL), instruction tuning, and prompting have recently been shown to improve the...
A key technology for the development of large language models (LLMs) involves instruction tuning tha...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
One of the most common solutions adopted by software researchers to address code generation is by tr...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Large language models (LLMs) have demonstrated significant potential in the realm of natural languag...
The automation of code review activities, a long-standing pursuit in software engineering, has been ...
Large language models~(LLMs) are instruction followers, but it can be challenging to find the best i...
Recently, Instruction fine-tuning has risen to prominence as a potential method for enhancing the ze...
Very large language models (LLMs), such as GPT-3 and Codex have achieved state-of-the-art performanc...
Recent work has shown that fine-tuning large pre-trained language models on a collection of tasks de...
Sparse Mixture-of-Experts (MoE) is a neural architecture design that can be utilized to add learnabl...
This paper systematically investigates the generation of code explanations by Large Language Models ...
Large Language Models (LLM) are a new class of computation engines, "programmed" via prompt engineer...
Large Language Models (LLMs) for code are a family of high-parameter, transformer-based neural netwo...
Multi-task learning (MTL), instruction tuning, and prompting have recently been shown to improve the...
A key technology for the development of large language models (LLMs) involves instruction tuning tha...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
One of the most common solutions adopted by software researchers to address code generation is by tr...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Large language models (LLMs) have demonstrated significant potential in the realm of natural languag...
The automation of code review activities, a long-standing pursuit in software engineering, has been ...