Cited 0 time in
Investigating Grammatical Transfer in Korean-English GPT2 Language Models
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Koo, Keonwoo | - |
| dc.contributor.author | Lee, Jaemin | - |
| dc.contributor.author | Park, Myung-Kwan | - |
| dc.date.accessioned | 2024-08-13T06:30:19Z | - |
| dc.date.available | 2024-08-13T06:30:19Z | - |
| dc.date.issued | 2024-06 | - |
| dc.identifier.issn | 1598-1398 | - |
| dc.identifier.issn | 2586-7474 | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/22862 | - |
| dc.description.abstract | With the recent success of artificial neural language model (LMs), their language acquisition has gained much attention (Futrell et al. 2019, Hu et al. 2020, Linzen et al. 2016, Warstadt et al. 2020, Wilcox et al. 2018). This paper delves into their second language (L2) acquisition, a largely unexplored area compared to their first language (L1) learning. The primary focus is on unraveling transfer effects originating from the L1’s linguistic structures. By closely examining our LMs’ performances on English grammar tasks, this study inspects how LMs encode abstract grammatical knowledge, particularly how pre-training biases acquired from Korean (L1) influence English (L2) performances in LMs. We present exploratory experiments where LMs were first trained on the dataset representing the initial language acquisition stage, followed by fine-tuning on the second language dataset. We analyzed cross-lingual transfer effects across diverse linguistic phenomena with the BLiMP test suite. We found that L1 pre-training did not accelerate linguistic generalization in the second language. Furthermore, our results revealed significant L1-interference, where the initial language knowledge hindered the LMs' ability to acquire and apply second language rules. © 2024 KASELL All rights reserved. | - |
| dc.format.extent | 21 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | 한국영어학회 | - |
| dc.title | Investigating Grammatical Transfer in Korean-English GPT2 Language Models | - |
| dc.type | Article | - |
| dc.publisher.location | 대한민국 | - |
| dc.identifier.doi | 10.15738/kjell.24..202406.568 | - |
| dc.identifier.scopusid | 2-s2.0-85197383955 | - |
| dc.identifier.bibliographicCitation | 영어학, v.24, pp 568 - 588 | - |
| dc.citation.title | 영어학 | - |
| dc.citation.volume | 24 | - |
| dc.citation.startPage | 568 | - |
| dc.citation.endPage | 588 | - |
| dc.type.docType | Article | - |
| dc.identifier.kciid | ART003105740 | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.description.journalRegisteredClass | kci | - |
| dc.subject.keywordAuthor | GPT-2 | - |
| dc.subject.keywordAuthor | L1-interference | - |
| dc.subject.keywordAuthor | neural language model | - |
| dc.subject.keywordAuthor | second language acquisition | - |
| dc.subject.keywordAuthor | transfer effects | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
