Assessing the Structural Profiles of L2-textbook Dataset on Transformer LMsAssessing the Structural Profiles of L2-textbook Dataset on Transformer LMs
- Other Titles
- Assessing the Structural Profiles of L2-textbook Dataset on Transformer LMs
- Authors
- 이재민; 박명관
- Issue Date
- Jan-2024
- Publisher
- 한국중원언어학회
- Keywords
- transformer language models; language acquisition; grammatical knowledge; textbook dataset; natural language process
- Citation
- 언어학 연구, no.70, pp 225 - 240
- Pages
- 16
- Indexed
- KCI
- Journal Title
- 언어학 연구
- Number
- 70
- Start Page
- 225
- End Page
- 240
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/21465
- DOI
- 10.17002/sil..70.20241.225
- ISSN
- 1975-8251
2508-4259
- Abstract
- Transformer-based language models (TLMs) employing multi-head self-attention methods have led to substantial enhancements in performance across diverse domains in Natural Language Processing (NLP). While current TLMs have demonstrated impressive capabilities by training on datasets hundreds of times larger than those akin to children's learning data, BabyBERTa has achieved meaningful performance by leveraging developmentally plausible datasets of comparable size. This study delves into the detailed evaluation of BabyBERTa's performance, aiming to gain insights into TLMs' language acquisition abilities and explore the feasibility of utilizing second language textbook dataset. Our analysis indicates that the dataset encompassing sentences with varied structures can effectively facilitate TLMs in acquiring grammatical knowledge.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Humanities > Division of English Language & Literature > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.