본문 바로가기
  • Home

Analysis of dynamic LSTM network activation in LSTM English learner language model: Subject-Verb Number Agreement task

  • Journal of The Korea Society of Computer and Information
  • Abbr : JKSCI
  • 2023, 28(6), pp.13-21
  • DOI : 10.9708/jksci.2023.28.06.013
  • Publisher : The Korean Society Of Computer And Information
  • Research Area : Engineering > Computer Science
  • Received : May 15, 2023
  • Accepted : June 5, 2023
  • Published : June 30, 2023

Euhee Kim 1

1신한대학교

Accredited

ABSTRACT

In this paper, we propose an approach to analyze the dynamic LSTM network activation in an LSTM English learner language model trained on a large corpus of English-speaking learners. The objective is to examine the relationship between network activation and performance on the Subject-Verb Number Agreement task. By employing a NA-task probing classifier and conducting ablation experiments, activation patterns are evaluated at each time step. The results reveal a strong link between network activation and the classifier's performance. The proposed model achieved 99.57% accuracy on the evaluation dataset for the NA-task, demonstrating its acquisition of correct grammar rules and accurate prediction ability. To analyze the influence of internal neurons on NA-task processing, specific LSTM neurons are removed and the model's performance is examined. Removing neuron 776 resulted in a more than 10% decrease in performance for plural subjects, while removing neuron 988 led to a 10% decrease in performance for singular subjects compared to the model before removal.

Citation status

* References for papers published after 2023 are currently being built.

This paper was written with support from the National Research Foundation of Korea.