Software correctness is crucial, with unit testing playing an indispensable role in the software development lifecycle. However, creating unit tests is time-consuming and costly, underlining the need for automation. Leveraging Large Language Models (LLMs) for unit test generation is a promising solution, but existing studies focus on simple, small-scale scenarios, leaving a gap in understanding LLMs' performance in real-world applications, particularly regarding integration and assessment efficacy at scale. Here, we present AgoneTest, a system focused on automatically generating and evaluating complex class-level test suites. Our contributions include a scalable automated system, a newly developed dataset for rigorous evaluation, and a detailed methodology for test quality assessment.

AgoneTest: Automated creation and assessment of Unit tests leveraging Large Language Models / Lops, Andrea; Narducci, Fedelucio; Ragone, Azzurra; Trizio, Michelantonio. - ELETTRONICO. - (2024), pp. 2440-2441. (Intervento presentato al convegno 39th ACM/IEEE International Conference on Automated Software Engineering, ASE 2024 tenutosi a usa nel 2024) [10.1145/3691620.3695318].

AgoneTest: Automated creation and assessment of Unit tests leveraging Large Language Models

Andrea Lops
;
Fedelucio Narducci;Azzurra Ragone;Michelantonio Trizio
2024-01-01

Abstract

Software correctness is crucial, with unit testing playing an indispensable role in the software development lifecycle. However, creating unit tests is time-consuming and costly, underlining the need for automation. Leveraging Large Language Models (LLMs) for unit test generation is a promising solution, but existing studies focus on simple, small-scale scenarios, leaving a gap in understanding LLMs' performance in real-world applications, particularly regarding integration and assessment efficacy at scale. Here, we present AgoneTest, a system focused on automatically generating and evaluating complex class-level test suites. Our contributions include a scalable automated system, a newly developed dataset for rigorous evaluation, and a detailed methodology for test quality assessment.
2024
39th ACM/IEEE International Conference on Automated Software Engineering, ASE 2024
AgoneTest: Automated creation and assessment of Unit tests leveraging Large Language Models / Lops, Andrea; Narducci, Fedelucio; Ragone, Azzurra; Trizio, Michelantonio. - ELETTRONICO. - (2024), pp. 2440-2441. (Intervento presentato al convegno 39th ACM/IEEE International Conference on Automated Software Engineering, ASE 2024 tenutosi a usa nel 2024) [10.1145/3691620.3695318].
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11589/284140
Citazioni
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
social impact