ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper Uncertainty-Driven Pessimistic Q-Ensemble for Offline-to-Online Reinforcement Learning
Cited - time in scopus Share share facebook twitter linkedin kakaostory
Authors
Ingook Jang, Seonghyun Kim
Issue Date
2022-12
Citation
Conference on Neural Information Processing Systems (NeurIPS) 2022 : Workshop, pp.1-5
Language
English
Type
Conference Paper
Abstract
Re-using existing offline reinforcement learning (RL) agents is an emerging topic for reducing the dominant computational cost for exploration in many settings. To effectively fine-tune the pre-trained offline policies, both offline samples and online interactions may be leveraged. In this paper, we propose the idea of incorporating a pessimistic Q-ensemble and an uncertainty quantification technique to effectively fine-tune offline agents. To stabilize online Q-function estimates during fine-tuning, the proposed method uses uncertainty estimation as a penalization for a replay buffer with a mixture of online interactions from the ensemble agent and offline samples from the behavioral policies. In various robotic tasks on D4RL benchmark, we show that our method outperforms the state-of-the-art algorithms in terms of the average return and the sample efficiency.
KSP Keywords
Fine-tuning, Q-Function, Reinforcement learning(RL), Uncertainty Quantification, computational cost, online interactions, state-of-The-Art, uncertainty estimation