본문 바로가기
  • Home

A Study on the Use of Large Language Models for Software Completeness Appraisal

  • Journal of Software Forensics
  • Abbr : JSAV
  • 2026, 22(1), pp.1~10
  • Publisher : Korea Software Assessment and Valuation Society
  • Research Area : Engineering > Computer Science
  • Received : February 23, 2026
  • Accepted : March 20, 2026
  • Published : March 31, 2026

Yukyong Kim 1

1숙명여자대학교

Accredited

ABSTRACT

Software completeness appraisal plays a critical role in contractual compliance verification, dispute resolution, and public procurement evaluation. It is a dynamic, execution-based assessment process that determines whether contracted requirements are fulfilled through observable system behavior. Large Language Models (LLMs) offer potential support for requirement understanding and code analysis, yet their susceptibility to hallucination and non-determinism limits their suitability as final decision-making tools in high-reliability contexts. This paper proposes a hallucination-controlled, LLM-assisted appraisal framework in which LLMs are restricted to auxiliary analytical roles. Specifically, LLMs support test scenario drafting, requirement-evidence semantic matching, and report drafting, while final completeness judgments are made by experts based on execution evidence and predefined rules. The proposed framework demonstrates how LLMs can be integrated into software completeness appraisal in a controlled manner, improving efficiency while preserving reliability.

Citation status

* References for papers published after 2024 are currently being built.