Integrating Large Language Models into Automated Software Testing
Yanet Sáez Iznaga (),
Luís Rato (),
Pedro Salgueiro and
Javier Lamar León
Additional contact information
Yanet Sáez Iznaga: Dectech, Rua Circular Norte do Parque Industrial e Tecnológico de Évora, Lote 2, 7005-841 Evora, Portugal
Luís Rato: VISTA Lab, ALGORITMI Research Center/LASI, University of Évora, 7000-671 Evora, Portugal
Pedro Salgueiro: VISTA Lab, ALGORITMI Research Center/LASI, University of Évora, 7000-671 Evora, Portugal
Javier Lamar León: VISTA Lab, ALGORITMI Research Center/LASI, University of Évora, 7000-671 Evora, Portugal
Future Internet, 2025, vol. 17, issue 10, 1-25
Abstract:
This work investigates the use of LLMs to enhance automation in software testing, with a particular focus on generating high-quality, context-aware test scripts from natural language descriptions, while addressing both text-to-code and text+code-to-code generation tasks. The Codestral Mamba model was fine-tuned by proposing a way to integrate LoRA matrices into its architecture, enabling efficient domain-specific adaptation and positioning Mamba as a viable alternative to Transformer-based models. The model was trained and evaluated on two benchmark datasets: CONCODE/CodeXGLUE and the proprietary TestCase2Code dataset. Through structured prompt engineering, the system was optimized to generate syntactically valid and semantically meaningful code for test cases. Experimental results demonstrate that the proposed methodology successfully enables the automatic generation of code-based test cases using large language models. In addition, this work reports secondary benefits, including improvements in test coverage, automation efficiency, and defect detection when compared to traditional manual approaches. The integration of LLMs into the software testing pipeline also showed potential for reducing time and cost while enhancing developer productivity and software quality. The findings suggest that LLM-driven approaches can be effectively aligned with continuous integration and deployment workflows. This work contributes to the growing body of research on AI-assisted software engineering and offers practical insights into the capabilities and limitations of current LLM technologies for testing automation.
Keywords: automated software testing; large language models; test case generation; low-rank adaptation codestral mamba model (search for similar items in EconPapers)
JEL-codes: O3 (search for similar items in EconPapers)
Date: 2025
References: View complete reference list from CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/1999-5903/17/10/476/pdf (application/pdf)
https://www.mdpi.com/1999-5903/17/10/476/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jftint:v:17:y:2025:i:10:p:476-:d:1774373
Access Statistics for this article
Future Internet is currently edited by Ms. Grace You
More articles in Future Internet from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().