ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Journal Article Prioritized Environment Configuration for Drone Control with Deep Reinforcement Learning
Cited 7 time in scopus Download 272 time Share share facebook twitter linkedin kakaostory
Authors
Sooyoung Jang, Changbeom Choi
Issue Date
2022-01
Citation
Human-centric Computing and Information Sciences, v.12, pp.1-16
ISSN
2192-1962
Publisher
KIPS-CSWRG
Language
English
Type
Journal Article
DOI
https://dx.doi.org/10.22967/HCIS.2022.12.002
Abstract
In reinforcement learning, first, the agent collects experiences by interacting with the environment through trial-and-errors (experience collection stage) and then learns from the collected experiences (learning stage). This two-stage training process repeats until the agent solves a given task and requires a lot of experience, computation power, and time for training the agent. Therefore, many studies are conducted to improve the training speed and performance to mitigate them, focusing on the learning stage. This paper focuses on the experience collection stage and proposes a prioritized environment configuration that prioritizes and stochastically samples the effective configuration for initializing the environment for every episode. Therefore, we can provide the environments initialized with the configuration suitable for effective experience collection to the agent. The proposed algorithm can complement the reinforcement learning algorithms that focus on the learning stage. We have shown speed and performance improvement by applying the prioritized environment configuration to an autonomous drone flight simulator. In addition, the results show that the proposed algorithm works well with both on-policy and off-policy reinforcement learning algorithms in distributed framework with multiple workers
KSP Keywords
Computation power, Deep reinforcement learning, Distributed framework, Effective configuration, Flight Simulator, Learning Stage, Off-policy reinforcement learning, Reinforcement Learning(RL), Training speed, Two-Stage, learning algorithms
This work is distributed under the term of Creative Commons License (CCL)
(CC BY NC)
CC BY NC