LLM4TDD: Best Practices for Test Driven Development Using Large Language Models
In today's society, we are becoming increasingly dependent on software systems. However, we also constantly witness the negative impacts of buggy software. Program synthesis aims to improve software correctness by automatically generating the program given an outline of the expected behavior. F...
Saved in:
Published in | 2024 IEEE/ACM International Workshop on Large Language Models for Code (LLM4Code) pp. 14 - 21 |
---|---|
Main Authors | , |
Format | Conference Proceeding |
Language | English |
Published |
ACM
20.04.2024
|
Subjects | |
Online Access | Get full text |
DOI | 10.1145/3643795.3648382 |
Cover
Summary: | In today's society, we are becoming increasingly dependent on software systems. However, we also constantly witness the negative impacts of buggy software. Program synthesis aims to improve software correctness by automatically generating the program given an outline of the expected behavior. For decades, program synthesis has been an active research field, with recent approaches looking to incorporate Large Language Model. This paper explores the concept of LLM4TDD, where we guide Large Language Models to generate code iteratively using a test-driven development methodology. We conduct an empirical evaluation using ChatGPT and coding problems from LeetCode to investigate the impact of different test, prompt and problem attributes on the efficacy of LLM4TDD.CCS CONCEPTS* Software and its engineering → Software development techniques. |
---|---|
DOI: | 10.1145/3643795.3648382 |