본문 바로가기
  • Home

Design and Development of Open-Source-Based Artificial Intelligence for Emotion Extraction from Voice

  • Journal of The Korea Society of Computer and Information
  • Abbr : JKSCI
  • 2024, 29(9), pp.79-87
  • DOI : 10.9708/jksci.2024.29.09.079
  • Publisher : The Korean Society Of Computer And Information
  • Research Area : Engineering > Computer Science
  • Received : August 12, 2024
  • Accepted : September 21, 2024
  • Published : September 30, 2024

Seong-Gun Yun 1 Hyeok-Chan Kwon 1 Eunju Park 1 Youngbok Cho 1

1안동대학교

Accredited

ABSTRACT

This study aims to improve communication for people with hearing impairments by developing artificial intelligence models that recognize and classify emotions from voice data. To achieve this, we utilized three major AI models: CNN-Transformer, HuBERT-Transformer, and Wav2Vec 2.0, to analyze users' voices in real-time and classify their emotions. To effectively extract features from voice data, we applied transformation techniques such as Mel-Frequency Cepstral Coefficient (MFCC), aiming to accurately capture the complex characteristics and subtle changes in emotions within the voice. Experimental results showed that the HuBERT-Transformer model demonstrated the highest accuracy, proving the effectiveness of combining pre-trained models and complex learning structures in the field of voice-based emotion recognition. This research presents the potential for advancements in emotion recognition technology using voice data and seeks new ways to improve communication and interaction for individuals with hearing impairments, marking its significance.

Citation status

* References for papers published after 2023 are currently being built.