Cited 7 time in
Dynamic Action Space Handling Method for Reinforcement Learning models
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Woo, Sangchul | - |
| dc.contributor.author | Sung, Yunsick | - |
| dc.date.accessioned | 2023-04-27T21:40:43Z | - |
| dc.date.available | 2023-04-27T21:40:43Z | - |
| dc.date.issued | 2020-10 | - |
| dc.identifier.issn | 1976-913X | - |
| dc.identifier.issn | 2092-805X | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/6087 | - |
| dc.description.abstract | Recently, extensive studies have been conducted to apply deep learning to reinforcement learning to solve the state-space problem. If the state-space problem was solved, reinforcement learning would become applicable in various fields. For example, users can utilize dance-tutorial systems to learn how to dance by watching and imitating a virtual instructor. The instructor can perform the optimal dance to the music, to which reinforcement learning is applied. In this study, we propose a method of reinforcement learning in which the action space is dynamically adjusted. Because actions that are not performed or are unlikely to be optimal are not learned, and the state space is not allocated, the learning time can be shortened, and the state space can be reduced. In an experiment, the proposed method shows results similar to those of traditional Q-learning even when the state space of the proposed method is reduced to approximately 0.33% of that of Q-learning. Consequently, the proposed method reduces the cost and time required for learning. Traditional Q-learning requires 6 million state spaces for learning 100,000 times. In contrast, the proposed method requires only 20,000 state spaces. A higher winning rate can be achieved in a shorter period of time by retrieving 20,000 state spaces instead of 6 million. | - |
| dc.format.extent | 8 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | KOREA INFORMATION PROCESSING SOC | - |
| dc.title | Dynamic Action Space Handling Method for Reinforcement Learning models | - |
| dc.type | Article | - |
| dc.publisher.location | 대한민국 | - |
| dc.identifier.doi | 10.3745/JIPS.02.0146 | - |
| dc.identifier.scopusid | 2-s2.0-85099188967 | - |
| dc.identifier.bibliographicCitation | JOURNAL OF INFORMATION PROCESSING SYSTEMS, v.16, no.5, pp 1223 - 1230 | - |
| dc.citation.title | JOURNAL OF INFORMATION PROCESSING SYSTEMS | - |
| dc.citation.volume | 16 | - |
| dc.citation.number | 5 | - |
| dc.citation.startPage | 1223 | - |
| dc.citation.endPage | 1230 | - |
| dc.type.docType | Article | - |
| dc.identifier.kciid | ART002642765 | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.description.journalRegisteredClass | esci | - |
| dc.description.journalRegisteredClass | kci | - |
| dc.relation.journalResearchArea | Computer Science | - |
| dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
| dc.subject.keywordAuthor | Dance Tutorial System | - |
| dc.subject.keywordAuthor | Q-Learning | - |
| dc.subject.keywordAuthor | Reinforcement Learning | - |
| dc.subject.keywordAuthor | Virtual Tutor | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
