Cited 0 time in
Evaluating L2 Training Methods in Neural Language Models
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | 이재민 | - |
| dc.contributor.author | 신정아 | - |
| dc.date.accessioned | 2025-01-15T06:30:17Z | - |
| dc.date.available | 2025-01-15T06:30:17Z | - |
| dc.date.issued | 2024-12 | - |
| dc.identifier.issn | 0254-4474 | - |
| dc.identifier.issn | 2586-7113 | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/56755 | - |
| dc.description.abstract | Recent advancements in language models (LMs) have significantly improved language processing capabilities; however, these models remain less efficient than human learning, especially when trained on developmentally plausible data volumes similar to those encountered by children (Warstadt & Bowman, 2022; Linzen, 2020). The inefficiency is even more pronounced in second language (L2) acquisition contexts, where cross-linguistic transfer is a key phenomenon (Papadimitriou & Jurafsky, 2020; Yadavalli et al., 2023). This study evaluates L2 training methods in neural language models by examining mutual L1-L2 influences during learning with developmentally plausible data volumes. We propose two approaches to mitigate catastrophic forgetting: the One-Stage Training (OST) method, which integrates L1 and L2 learning into a single stage, and the One-Stage Mixed Training (OSMT) method, which refines OST by incorporating L1 data into the L2 stage for more realistic simulation of bilingual learning. Through continuous syntactic evaluations throughout training, we analyzed how L1 performance changes during L2 acquisition and how cross-linguistics transfer emerges in Korean and English. The results indicate that OST and OSMT effectively mitigated catastrophic forgetting and supported more stable learning compared to the conventional Two-Stage Training method. OSMT achieved superior integration of L1 and L2 structures while revealing negative transfer effects from Korean (L1) to English (L2). These findings provide valuable insights into both neural model training and human-like L2 acquisition processes. | - |
| dc.format.extent | 23 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | 서울대학교 언어교육원 | - |
| dc.title | Evaluating L2 Training Methods in Neural Language Models | - |
| dc.type | Article | - |
| dc.publisher.location | 대한민국 | - |
| dc.identifier.doi | 10.30961/lr.2024.60.3.323 | - |
| dc.identifier.bibliographicCitation | 어학연구, v.60, no.3, pp 323 - 345 | - |
| dc.citation.title | 어학연구 | - |
| dc.citation.volume | 60 | - |
| dc.citation.number | 3 | - |
| dc.citation.startPage | 323 | - |
| dc.citation.endPage | 345 | - |
| dc.identifier.kciid | ART003158471 | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | kci | - |
| dc.subject.keywordAuthor | developmentally plausible data | - |
| dc.subject.keywordAuthor | cross-linguistic transfer | - |
| dc.subject.keywordAuthor | second language acquisition | - |
| dc.subject.keywordAuthor | neural language models | - |
| dc.subject.keywordAuthor | L2 language models | - |
| dc.subject.keywordAuthor | catastrophic forgetting | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
