Asahina, Wakana, Iwamoto, Naoya, Shum, Hubert P. H. and Morishima, Shigeo (2016) Automatic Dance Generation System Considering Sign Language Information. In: SIGGRAPH 2016 - 43rd International Conference and Exhibition on Computer Graphics and Interactive Techniques, 24th - 28th July 2016, Anaheim, California.
|
Text (Full text)
siggraph2016[1].pdf - Published Version Download (548kB) | Preview |
Abstract
In recent years, thanks to the development of 3DCG animation editing tools (e.g. MikuMikuDance), a lot of 3D character dance animation movies are created by amateur users. However it is very difficult to create choreography from scratch without any technical knowledge. Shiratori et al. [2006] produced the dance automatic generation system considering rhythm and intensity of dance motions. However each segment is selected randomly from database, so the generated dance motion has no linguistic or emotional meanings. Takano et al. [2010] produced a human motion generation system considering motion labels. However they use simple motion labels like “running” or “jump”, so they cannot generate motions that express emotions. In reality, professional dancers make choreography based on music features or lyrics in music, and express emotion or how they feel in music. In our work, we aim at generating more emotional dance motion easily. Therefore, we use linguistic information in lyrics, and generate dance motion.
In this paper, we propose the system to generate the sign dance motion from continuous sign language motion based on lyrics of music. This system could help the deaf to listen to music as visualized music application.
Item Type: | Conference or Workshop Item (Paper) |
---|---|
Subjects: | G400 Computer Science |
Department: | Faculties > Engineering and Environment > Computer and Information Sciences |
Depositing User: | Hubert Shum |
Date Deposited: | 02 Nov 2016 10:20 |
Last Modified: | 01 Aug 2021 12:47 |
URI: | http://nrl.northumbria.ac.uk/id/eprint/28248 |
Downloads
Downloads per month over past year