Reactive virtual agent learning for NUI-based HRI applications
- Authors
- Jin, Daxing; Cho, Seoungjae; Sung, Yunsick; Cho, Kyungeun; Um, Kyhyun
- Issue Date
- Dec-2016
- Publisher
- SPRINGER
- Keywords
- Natural user interface; Natural user experience; Human-robot interaction; Virtual agent learning
- Citation
- MULTIMEDIA TOOLS AND APPLICATIONS, v.75, no.23, pp 15157 - 15170
- Pages
- 14
- Indexed
- SCIE
SCOPUS
- Journal Title
- MULTIMEDIA TOOLS AND APPLICATIONS
- Volume
- 75
- Number
- 23
- Start Page
- 15157
- End Page
- 15170
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/23455
- DOI
- 10.1007/s11042-014-2048-5
- ISSN
- 1380-7501
1573-7721
- Abstract
- The natural user interface (NUI) has been investigated in a variety of fields in application software. This paper proposes an approach to generate virtual agents that can support users for NUI-based applications through human-robot interaction (HRI) learning in a virtual environment. Conventional human-robot interaction (HRI) learning is carried out by repeating processes that are time-consuming, complicated and dangerous because of certain features of robots. Therefore, a method is needed to train virtual agents that interact with virtual humans imitating human movements in a virtual environment. Then the result of this virtual agent can be applied to NUI-based interactive applications after the interaction learning is completed. The proposed method was applied to a model of a typical house in virtual environment with virtual human performing daily-life activities such as washing, eating, and watching TV. The results show that the virtual agent can predict a human's intent, identify actions that are helpful to the human, and can provide services 16 % faster than a virtual agent trained using traditional Q-learning.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.