ETRI-Knowledge Sharing Plaform



논문 검색
구분 SCI
연도 ~ 키워드


학술대회 Robots Learn Social Skills: End-to-End Learning of Co-Speech Gesture Generation for Humanoid Robots
Cited 145 time in scopus Download 16 time Share share facebook twitter linkedin kakaostory
윤영우, 고우리, 장민수, 이재연, 김재홍, 이기혁
International Conference on Robotics and Automation (ICRA) 2019, pp.4303-4309
18HS4800, 고령 사회에 대응하기 위한 실환경 휴먼케어 로봇 기술 개발, 이재연
Co-speech gestures enhance interaction experiences between humans as well as between humans and robots. Most existing robots use rule-based speech-gesture association, but this requires human labor and prior knowledge of experts to be implemented. We present a learning-based co-speech gesture generation that is learned from 52 h of TED talks. The proposed end-to-end neural network model consists of an encoder for speech text understanding and a decoder to generate a sequence of gestures. The model successfully produces various gestures including iconic, metaphoric, deictic, and beat gestures. In a subjective evaluation, participants reported that the gestures were human-like and matched the speech content. We also demonstrate a co-speech gesture with a NAO robot working in real time.
KSP 제안 키워드
2 H, End to End(E2E), Gesture generation, Human labor, Human-like, Learning-based, Nao robot, Real-Time, Rule-based, TED talks, Text understanding