The Impact of Prompt Programming on Function-Level Code Generation
Artikel i vetenskaplig tidskrift, 2025

Large Language Models (LLMs) are increasingly used by software engineers for code generation. However, limitations of LLMs such as irrelevant or incorrect code have highlighted the need for prompt programming (or prompt engineering) where engineers apply specific prompt techniques (e.g., chain-of-thought or input-output examples) to improve the generated code. While some prompt techniques have been studied, the impact of different techniques—and their interactions— on code generation is still not fully understood. In this study, we introduce CodePromptEval, a dataset of 7072 prompts designed to evaluate five prompt techniques (few-shot, persona, chain-of-thought, function signature, list of packages) and their effect on the correctness, similarity, and quality of complete functions generated by three LLMs (GPT-4o, Llama3, and Mistral). Our findings show that while certain prompt techniques significantly influence the generated code, combining multiple techniques does not necessarily improve the outcome. Additionally, we observed a trade-off between correctness and quality when using prompt techniques. Our dataset and replication package enable future research on improving LLM-generated code and evaluating new prompt techniques.

Författare

Ranim Khojah

Chalmers, Data- och informationsteknik, Interaktionsdesign och Software Engineering

Francisco Gomes

Chalmers, Data- och informationsteknik, Interaktionsdesign och Software Engineering

Mazen Mohamad

Chalmers, Data- och informationsteknik, Interaktionsdesign och Software Engineering

RISE Research Institutes of Sweden

Philipp Leitner

Chalmers, Data- och informationsteknik, Interaktionsdesign och Software Engineering

IEEE Transactions on Software Engineering

0098-5589 (ISSN) 19393520 (eISSN)

Vol. 51 8 2381-2395

Ämneskategorier (SSIF 2025)

Programvaruteknik

DOI

10.1109/TSE.2025.3587794

Mer information

Senast uppdaterat

2025-09-03