Skip to main navigation Skip to search Skip to main content

Prompt Tuning in Code Intelligence: An Experimental Evaluation

  • Chaozheng Wang
  • , Yuanhang Yang
  • , Cuiyun Gao*
  • , Yun Peng
  • , Hongyu Zhang
  • , Michael R. Lyu
  • *Corresponding author for this work
  • Harbin Institute of Technology Shenzhen
  • Chinese University of Hong Kong
  • University of Newcastle
  • Chongqing University

Research output: Contribution to journalArticlepeer-review

Abstract

Pre-trained models have been shown effective in many code intelligence tasks, such as automatic code summarization and defect prediction. These models are pre-trained on large-scale unlabeled corpus and then fine-tuned in downstream tasks. However, as the inputs to pre-training and downstream tasks are in different forms, it is hard to fully explore the knowledge of pre-trained models. Besides, the performance of fine-tuning strongly relies on the amount of downstream task data, while in practice, the data scarcity scenarios are common. Recent studies in the natural language processing (NLP) field show that prompt tuning, a new paradigm for tuning, alleviates the above issues and achieves promising results in various NLP tasks. In prompt tuning, the prompts inserted during tuning provide task-specific knowledge, which is especially beneficial for tasks with relatively scarce data. In this article, we empirically evaluate the usage and effect of prompt tuning in code intelligence tasks. We conduct prompt tuning on popular pre-trained models CodeBERT and CodeT5 and experiment with four code intelligence tasks including defect prediction, code search, code summarization, and code translation. Our experimental results show that prompt tuning consistently outperforms fine-tuning in all four tasks. In addition, prompt tuning shows great potential in low-resource scenarios, e.g., improving the BLEU scores of fine-tuning by more than 26% on average for code summarization. Our results suggest that instead of fine-tuning, we could adapt prompt tuning for code intelligence tasks to achieve better performance, especially when lacking task-specific data. We also discuss the implications for adapting prompt tuning in code intelligence tasks.

Original languageEnglish
Pages (from-to)4869-4885
Number of pages17
JournalIEEE Transactions on Software Engineering
Volume49
Issue number11
DOIs
StatePublished - 1 Nov 2023
Externally publishedYes

Keywords

  • Code intelligence
  • empirical study
  • prompt tuning

Fingerprint

Dive into the research topics of 'Prompt Tuning in Code Intelligence: An Experimental Evaluation'. Together they form a unique fingerprint.

Cite this