ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper Coupling Vision and Proprioception for Sample-Efficient, Object-Occlusion-Robust Robotic Manipulation
Cited - time in scopus Share share facebook twitter linkedin kakaostory
Authors
Samyeul Noh, Hyun Myung
Issue Date
2023-10
Citation
International Conference on Computer Vision Workshops (ICCVW) 2023, pp.1-9
Language
English
Type
Conference Paper
Abstract
Recent advances in visual reinforcement learning (visual RL), which learns directly from image pixels, have bridged the gap between state-based and image-based training. These advances have been achieved through the effective combination of traditional RL algorithms with latent representation learning. However, visual RL still struggles with solving robotic manipulation tasks involving object occlusions; for example, lifting an object occluded by an obstacle. In this paper, we propose a multimodal RL method that is sample-efficient and robust to object occlusions by effectively coupling two different types of raw sensory data: vision and proprioception. Our method is able to jointly learn multimodal latent representations and a policy in a fully end-to-end manner from vision and proprioception without the need for pre-training. We show that our method outperforms both current state-of-the-art visual RL and statebased RL methods on robotic manipulation tasks involving object occlusions in terms of sample efficiency and task performance, without any prior knowledge such as pre-defined coordinate states or pre-trained representations.
KSP Keywords
Current state, End to End(E2E), Image-based, Latent representations, Pre-Training, Reinforcement Learning(RL), Representation learning, image pixels, need for, prior knowledge, robotic manipulation