Can an L2-neural LM Generalize Filler-gap Dependency?
- Authors
- 최선주; 윤영도; 박명관
- Issue Date
- Nov-2022
- Publisher
- 한국현대언어학회
- Keywords
- filler-gap dependency; neural language model; syntactic priming; adaptation
- Citation
- 언어연구, v.38, no.3, pp 323 - 337
- Pages
- 15
- Indexed
- KCI
- Journal Title
- 언어연구
- Volume
- 38
- Number
- 3
- Start Page
- 323
- End Page
- 337
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/2250
- DOI
- 10.18627/jslg.38.3.202211.323
- ISSN
- 1225-4770
2671-6151
- Abstract
- Recent studies have shown
that recurrent neural language models (LMs) can understand sentences involving
filler-gap dependency (Chowdhury & Zamparelli, 2018; Wilcox et al., 2018,
2019). However, their behavior does not encode the underlying constraints that
govern filler-gap acceptability. In this vein, significant issues remain about the
extent to which LMs acquire specific linguistic constructions and whether these
models recognize an abstract property of syntax in their representations. In this
paper, following the lead of Bhattacharya and van Schijndel (2020), we further
test whether the L2 neural LM can learn abstract syntactic constraints that have
been claimed to govern the behavior of filler-gap constructions. To see this, we
implement the L2 neural LM trained on the L2 corpus of English textbooks
published in Korea for the last two decades, and then we test the representational
overlap between disparate filler-gap constructions based on the syntactic priming
paradigm. Unlike the previous studies of L1-neural LMs, we could not find
sufficient evidence showing that the L2 neural LM learns a general representation
of the existence of filler-gap dependency and the shared underlying constraints.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - Dharma College > 1. Journal Articles
- College of Humanities > Division of English Language & Literature > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.