ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper Time-Varying Constraint-Aware Reinforcement Learning for Energy Storage Control
Cited - time in scopus Share share facebook twitter linkedin kakaostory
Authors
Jaeik Jeong, Tai-Yeon Ku, Wan-Ki Park
Issue Date
2024-05
Citation
International Conference on Learning Representations (ICLR) 2024, pp.1-6
Language
English
Type
Conference Paper
Abstract
Energy storage devices, such as batteries, thermal energy storages, and hydrogen systems, can help mitigate climate change by ensuring a more stable and sustainable power supply. To maximize the effectiveness of such energy storage, determining the appropriate charging and discharging amounts for each time period is crucial. Reinforcement learning is preferred over traditional optimization for the control of energy storage due to its ability to adapt to dynamic and complex environments. However, the continuous nature of charging and discharging levels in energy storage poses limitations for discrete reinforcement learning, and time-varying feasible charge-discharge range based on state of charge (SoC) variability also limits the conventional continuous reinforcement learning. In this paper, we propose a continuous reinforcement learning approach that takes into account the time-varying feasible charge-discharge range. An additional objective function was introduced for learning the feasible action range for each time period, supplementing the objectives of training the actor for policy learning and the critic for value learning. This actively promotes the utilization of energy storage by preventing them from getting stuck in suboptimal states, such as continuous full charging or discharging. This is achieved through the enforcement of the charging and discharging levels into the feasible action range. The experimental results demonstrated that the proposed method further maximized the effectiveness of energy storage by actively enhancing its utilization.
KSP Keywords
Climate Change, Constraint-aware, Control energy, Hydrogen systems, Learning approach, Policy learning, Reinforcement Learning(RL), Sustainable power, Time period, charge-discharge, charging and discharging