The field of software development and testing is witnessing significant advancements with the integration of Large Language Models (LLMs). Recent research has focused on improving the capabilities of LLMs in generating high-quality code, testing software applications, and evaluating their performance. One notable direction is the development of novel frameworks and methods for automated testing and evaluation of LLMs, such as interactive evaluation frameworks and guideline-upholding tests. Another area of research is the application of LLMs in software development, including code generation, code completion, and code review. Furthermore, researchers are exploring the use of LLMs in identifying and addressing potential biases and flaws in software applications. Noteworthy papers in this area include SATORI, which introduces a static test oracle generation approach for REST APIs, and LaQual, which proposes a framework for automated evaluation of LLM app quality. Overall, these advancements have the potential to significantly improve the efficiency, effectiveness, and reliability of software development and testing processes.
Advancements in Large Language Models for Software Development and Testing
Sources
Research on intelligent generation of structural demolition suggestions based on multi-model collaboration
Functional Consistency of LLM Code Embeddings: A Self-Evolving Data Synthesis Framework for Benchmarking