Deep Q-network-based multi-criteria decision-making framework for virtual simulation environment
- Authors
- Jang, Hyeonjun; Hao, Shujia; Chu, Phuong Minh; Sharma, Pradip Kumar; Sung, Yunsick; Cho, Kyungeun
- Issue Date
- Sep-2021
- Publisher
- SPRINGER LONDON LTD
- Keywords
- Deep learning; Big data; Motivation system; Behavior planning; Nature inspired algorithm
- Citation
- NEURAL COMPUTING & APPLICATIONS, v.33, no.17, pp 10657 - 10671
- Pages
- 15
- Indexed
- SCIE
SCOPUS
- Journal Title
- NEURAL COMPUTING & APPLICATIONS
- Volume
- 33
- Number
- 17
- Start Page
- 10657
- End Page
- 10671
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/4502
- DOI
- 10.1007/s00521-020-04918-3
- ISSN
- 0941-0643
1433-3058
- Abstract
- Deep learning improves the realistic expression of virtual simulations specifically to solve multi-criteria decision-making problems, which are generally rely on high-performance artificial intelligence. This study was inspired by the motivation theory and natural life observations. Recently, motivation-based control has been actively studied for realistic expression, but it presents various problems. For instance, it is hard to define the relation among multiple motivations and to select goals based on multiple motivations. Behaviors should generally be practiced to take into account motivations and goals. This paper proposes a deep Q-network (DQN)-based multi-criteria decision-making framework for virtual agents in real time to automatically select goals based on motivations in virtual simulation environments and to plan relevant behaviors to achieve those goals. All motivations are classified according to the five-level Maslow's hierarchy of needs, and the virtual agents train a double DQN by big social data, select optimal goals depending on motivations, and perform behaviors relying on a predefined hierarchical task networks (HTNs). Compared to the state-of-the-art method, the proposed framework is efficient and reduced the average loss from 0.1239 to 0.0491 and increased accuracy from 63.24 to 80.15%. For behavioral performance using predefined HTNs, the number of methods has increased from 35 in the Q network to 1511 in the proposed framework, and the computation time of 10,000 behavior plans reduced from 0.118 to 0.1079 s.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.