Cited 0 time in
Probing Good-Enough Processing in Large Language Models with a Paraphrasing Task
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Jonghyun Lee | - |
| dc.contributor.author | Jeong-Ah Shin | - |
| dc.date.accessioned | 2026-02-27T18:00:33Z | - |
| dc.date.available | 2026-02-27T18:00:33Z | - |
| dc.date.issued | 2026-01 | - |
| dc.identifier.issn | 1598-1398 | - |
| dc.identifier.issn | 2586-7474 | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/63821 | - |
| dc.description.abstract | This study investigates whether large language models (LLMs) exhibit human-like ‘good-enough’ processing patterns in syntactic comprehension or demonstrate mechanical accuracy. Previous research using forced-choice question-answering paradigms revealed that LLMs display incomplete syntactic reanalysis similar to humans when processing garden-path sentences. However, concerns arose that these patterns might reflect methodological artifacts rather than genuine processing characteristics, as direct questioning could bias models toward initial misinterpretations. To address this limitation, we employed a paraphrasing task that requires comprehensive sentence reformulation rather than binary responses, following Patson et al. (2009). We tested GPT-3.5 and GPT-4 on 24 garden-path sentences containing Optionally Transitive (OT) and Reflexive Absolute Transitive (RAT) verbs. Results demonstrate that good-enough processing patterns persist across both paradigms, with LLMs continuing to exhibit partial reanalysis in garden-path conditions even when generating full paraphrases. This confirms that previously observed error patterns represent genuine syntactic processing characteristics rather than experimental artifacts. Notably, GPT-4 showed improved performance in the paraphrasing task compared to forced-choice experiments, suggesting task-dependent variation in processing depth. Both models exhibited human-like incomplete processing despite their substantial computational resources, indicating that their pattern-matching mechanisms favor processing shortcuts over complete syntactic interpretation. These findings reveal that LLMs demonstrate good-enough processing similar to humans, with performance varying systematically across task formats. | - |
| dc.format.extent | 15 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | 한국영어학회 | - |
| dc.title | Probing Good-Enough Processing in Large Language Models with a Paraphrasing Task | - |
| dc.type | Article | - |
| dc.publisher.location | 대한민국 | - |
| dc.identifier.doi | 10.15738/kjell.26..202601.127 | - |
| dc.identifier.scopusid | 2-s2.0-105030446730 | - |
| dc.identifier.bibliographicCitation | 영어학, v.26, pp 127 - 141 | - |
| dc.citation.title | 영어학 | - |
| dc.citation.volume | 26 | - |
| dc.citation.startPage | 127 | - |
| dc.citation.endPage | 141 | - |
| dc.type.docType | Y | - |
| dc.identifier.kciid | ART003301434 | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.description.journalRegisteredClass | kci | - |
| dc.subject.keywordAuthor | large language models | - |
| dc.subject.keywordAuthor | garden-path sentences | - |
| dc.subject.keywordAuthor | good-enough processing | - |
| dc.subject.keywordAuthor | syntactic processing | - |
| dc.subject.keywordAuthor | paraphrasing task | - |
| dc.subject.keywordAuthor | ChatGPT | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
