본문 바로가기
  • Home

Empirical Study on 70B-Level AI Interview Evaluation Performance of Small LLaMA 3.1 8B Using Generative AI and LLM Techniques

  • Journal of Internet of Things and Convergence
  • Abbr : JKIOTS
  • 2025, 11(6), 28
  • Publisher : The Korea Internet of Things Society
  • Research Area : Engineering > Computer Science > Internet Information Processing
  • Received : November 14, 2025
  • Accepted : December 21, 2025
  • Published : December 31, 2025

Ji Soo Ryu 1 Jung, Soon Ki 2

1경북대학교 컴퓨터학부 대학원
2경북대학교

Accredited

ABSTRACT

This study proposes an integrated AI interviewer system framework for on-premise environments. The system combines an STT–LLM pipeline, a Rasa-based dialogue manager, and a KoBERT–MeCab preprocessing module to enable real-time interview dialogue processing, automated evaluation, and feedback refinement. Applicants’ spoken responses are converted via STT, after which the LLM functions as an interviewer and evaluates responses based on professional relevance, logical consistency, coherence, and feedback usefulness. The application of Retrieval-Augmented Generation (RAG) and Reflection Tuning improves response quality and evaluation consistency. Experimental results show that the LLaMA 3.1 8B model with the proposed architecture demonstrates competitive performance compared to the 70B model under identical conditions, supporting its applicability from a non-inferiority perspective. These findings suggest that small-scale language models can achieve performance comparable to larger models through appropriate architectural design.

Citation status

* References for papers published after 2024 are currently being built.