Cited 0 time in
How are Korean Neural Language Models ‘surprised’ Layerwisely?
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | 최선주 | - |
| dc.contributor.author | 박명관 | - |
| dc.contributor.author | 김유희 | - |
| dc.date.accessioned | 2023-04-27T15:40:23Z | - |
| dc.date.available | 2023-04-27T15:40:23Z | - |
| dc.date.issued | 2021-11 | - |
| dc.identifier.issn | 1225-2522 | - |
| dc.identifier.issn | 2508-4267 | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/4213 | - |
| dc.description.abstract | Since the introduction of BERT, recent works have shown success in detecting when a word is anomalous given sentence context. Since likelihood score is not an appropriate tool in identifying the exact property of linguistic anomaly, Li et al. (2021) recently adopt Gaussian models for density estimation at intermediate layers of pretrained language models. They find that different English pretrained language models employ separate mechanisms to recognize different types of linguistic anomaly. In keeping with Li et al.‘s methodology, we probe whether Korean counterparts such as KoBERT and KR-BERT are sensitive to different levels of linguistic anomaly, just as English-based language models are. To investigate the issue concerned, we construct an experiment with a suite of test data involving morphosyntactic, semantic, and commonsense anomaly in Korean and apply the two Korean-based models to test relevant sentences. We find that KoBERT and KR-BERT show relatively higher surprisal gaps throughout layers when the anomaly is morphosyntactic than when the anomaly is semantic. By contrast, commonsense anomaly does not exhibit any surprisal gap in any layer. We thus report that, like their English counterparts, KoBERT and KR-BERT use different mechanisms to track the different types of linguistic anomaly. | - |
| dc.format.extent | 17 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | 한국언어과학회 | - |
| dc.title | How are Korean Neural Language Models ‘surprised’ Layerwisely? | - |
| dc.title.alternative | How are Korean Neural Language Models ‘surprised’ Layerwisely? | - |
| dc.type | Article | - |
| dc.publisher.location | 대한민국 | - |
| dc.identifier.doi | 10.14384/kals.2021.28.4.301 | - |
| dc.identifier.bibliographicCitation | 언어과학, v.28, no.4, pp 301 - 317 | - |
| dc.citation.title | 언어과학 | - |
| dc.citation.volume | 28 | - |
| dc.citation.number | 4 | - |
| dc.citation.startPage | 301 | - |
| dc.citation.endPage | 317 | - |
| dc.identifier.kciid | ART002777986 | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | kci | - |
| dc.subject.keywordAuthor | KR-BERT | - |
| dc.subject.keywordAuthor | KoBERT | - |
| dc.subject.keywordAuthor | linguistic anomaly | - |
| dc.subject.keywordAuthor | surprisal gap | - |
| dc.subject.keywordAuthor | layerwise | - |
| dc.subject.keywordAuthor | 한국어 신경망 언어모델 | - |
| dc.subject.keywordAuthor | 언어학적 변칙 | - |
| dc.subject.keywordAuthor | ‘놀라움’ 차이 | - |
| dc.subject.keywordAuthor | 신경망 층별 분석 | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
