ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Journal Article Exploring the Preference for Discrete over Continuous Reinforcement Learning in Energy Storage Arbitrage
Cited 0 time in scopus Download 85 time Share share facebook twitter linkedin kakaostory
Authors
Jaeik Jeong, Tai-Yeon Ku, Wan-Ki Park
Issue Date
2024-12
Citation
Energies, v.17, no.23, pp.1-17
ISSN
1996-1073
Publisher
Multidisciplinary Digital Publishing Institute (MDPI)
Language
English
Type
Journal Article
DOI
https://dx.doi.org/10.3390/en17235876
Abstract
In recent research addressing energy arbitrage with energy storage systems (ESSs), discrete reinforcement learning (RL) has often been employed, while the underlying reasons for this preference have not been explicitly clarified. This paper aims to elucidate why discrete RL tends to be more suitable than continuous RL for energy arbitrage problems. When using continuous RL, the charging and discharging actions determined by the agent often exceed the physical limits of the ESS, necessitating clipping to the boundary values. This introduces a critical issue where the learned actions become stuck at the state of charge (SoC) boundaries, hindering effective learning. Although recent advancements in constrained RL offer potential solutions, their application often results in overly conservative policies, preventing the full utilization of ESS capabilities. In contrast, discrete RL, while lacking in granular control, successfully avoids these two key challenges, as demonstrated by simulation results showing superior performance. Additionally, it was found that, due to its characteristics, discrete RL more easily drives the ESS towards fully charged or fully discharged states, thereby increasing the utilization of the storage system. Our findings provide a solid justification for the prevalent use of discrete RL in recent studies involving energy arbitrage with ESSs, offering new insights into the strategic selection of RL methods in this domain. Looking ahead, improving performance will require further advancements in continuous RL methods. This study provides valuable direction for future research in continuous RL, highlighting the challenges and potential strategies to overcome them to fully exploit ESS capabilities.
KSP Keywords
Energy Storage Systems(ESS), Energy arbitrage, Energy storage(ES), Granular control, Potential solutions, Reinforcement learning(RL), charging and discharging, physical limits, simulation results, superior performance
This work is distributed under the term of Creative Commons License (CCL)
(CC BY)
CC BY