Cited 0 time in
A Reinforcement Learning Framework for Personalized Anticoagulation Dosing in Critical Care: Integrating Batch-Constrained Policy Optimization and Off-Policy Evaluation
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Lim, Yooseok | - |
| dc.contributor.author | Park, In-Beom | - |
| dc.contributor.author | Lee, Sujee | - |
| dc.date.accessioned | 2025-12-10T03:00:32Z | - |
| dc.date.available | 2025-12-10T03:00:32Z | - |
| dc.date.issued | 2025 | - |
| dc.identifier.issn | 2169-3536 | - |
| dc.identifier.issn | 2169-3536 | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/62244 | - |
| dc.description.abstract | Precise medication dosing in the intensive care unit (ICU) is vital for patient survival. Heparin, a widely used anticoagulant, requires careful administration due to patient-specific variability, and inappropriate dosing can cause severe complications such as stroke or hemorrhage. This study introduces a reinforcement learning (RL)-based decision-support framework for heparin dosing, integrating offline RL algorithms with rigorous evaluation. We employ Batch-Constrained Q-Learning (BCQ) to learn an optimal dosing policy from retrospective data, addressing distributional shift inherent in offline settings. The dosing policies are trained on the MIMIC-III database and evaluated on the MIMIC-IV database, and vice versa. Policy effectiveness is assessed through multiple off-policy evaluation (OPE) methods, demonstrating higher expected returns than clinician-derived strategies. Interpretability is enhanced through t-SNE visualization, showing that Q-values are well aligned with therapeutic aPTT targets. To our knowledge, this is the first study to combine BCQ, multi-metric OPE, and interpretability analysis for anticoagulation management across two large-scale ICU cohorts. By advancing both methodological rigor and clinical relevance, this work provides a foundation for reliable RL-based decision-support systems in critical care. © 2013 IEEE. | - |
| dc.format.extent | 13 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | IEEE | - |
| dc.title | A Reinforcement Learning Framework for Personalized Anticoagulation Dosing in Critical Care: Integrating Batch-Constrained Policy Optimization and Off-Policy Evaluation | - |
| dc.type | Article | - |
| dc.publisher.location | 미국 | - |
| dc.identifier.doi | 10.1109/ACCESS.2025.3638417 | - |
| dc.identifier.scopusid | 2-s2.0-105023313047 | - |
| dc.identifier.wosid | 001631912500036 | - |
| dc.identifier.bibliographicCitation | IEEE Access, v.13, pp 203145 - 203157 | - |
| dc.citation.title | IEEE Access | - |
| dc.citation.volume | 13 | - |
| dc.citation.startPage | 203145 | - |
| dc.citation.endPage | 203157 | - |
| dc.type.docType | Article | - |
| dc.description.isOpenAccess | Y | - |
| dc.description.journalRegisteredClass | scie | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.relation.journalResearchArea | Computer Science | - |
| dc.relation.journalResearchArea | Engineering | - |
| dc.relation.journalResearchArea | Telecommunications | - |
| dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
| dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
| dc.relation.journalWebOfScienceCategory | Telecommunications | - |
| dc.subject.keywordAuthor | Batch-Constrained Policy | - |
| dc.subject.keywordAuthor | Medical Information Mart for Intensive Care | - |
| dc.subject.keywordAuthor | Off-Policy Evaluation | - |
| dc.subject.keywordAuthor | Personalized Heparin Dosing Policy | - |
| dc.subject.keywordAuthor | Reinforcement Learning | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
