Detailed Information

Cited 1 time in webofscience Cited 1 time in scopus
Metadata Downloads

Data generation and representation method for 3D video conferencing using programming by demonstration

Authors
Sung, YunsickCho, Kyungeun
Issue Date
Nov-2013
Publisher
SPRINGER
Keywords
Video conferencing; Programming by demonstration; Maximin Selection Algorithm; Virtual human; Behavior generation
Citation
MULTIMEDIA TOOLS AND APPLICATIONS, v.67, no.1, pp 71 - 95
Pages
25
Indexed
SCIE
SCOPUS
Journal Title
MULTIMEDIA TOOLS AND APPLICATIONS
Volume
67
Number
1
Start Page
71
End Page
95
URI
https://scholarworks.dongguk.edu/handle/sw.dongguk/17594
DOI
10.1007/s11042-011-0941-8
ISSN
1380-7501
1573-7721
Abstract
Video conferencing provides an environment for multiple users linked on a network to have meetings. Since a large quantity of audio and video data are transferred to multiple users in real time, research into reducing the quantity of data to be transferred has been drawing attention. Such methods extract and transfer only the features of a user from video data and then reconstruct a video conference using virtual humans. The disadvantage of such an approach is that only the positions and features of hands and heads are extracted and reconstructed, whilst the other virtual body parts do not follow the user. In order to enable a virtual human to accurately mimic the entire body of the user in a 3D virtual conference, we examined what features should be extracted to express a user more clearly and how they can be reproduced by a virtual human. This 3D video conferencing estimates the user's pose by comparing predefined images with a photographed user's image and generates a virtual human that takes the estimated pose. However, this requires predefining a diverse set of images for pose estimation and, moreover, it is difficult to define behaviors that can express poses correctly. This paper proposes a framework to automatically generate the pose-images used to estimate a user's pose and the behaviors required to present a user using a virtual human in a 3D video conference. The method for applying this framework to a 3D video conference on the basis of the automatically generated data is also described. In the experiment, the framework proposed in this paper was implemented in a mobile device. The generation process of poses and behaviors of virtual human was verified. Finally, by applying programming by demonstration, we developed a system that can automatically collect the various data necessary for a video conference directly without any prior knowledge of the video conference system.
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Sung, Yunsick photo

Sung, Yunsick
College of Advanced Convergence Engineering (Department of Computer Science and Artificial Intelligence)
Read more

Altmetrics

Total Views & Downloads

BROWSE