ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper TTL-Based Cache Utility Maximization Using Deep Reinforcement Learning
Cited 4 time in scopus Share share facebook twitter linkedin kakaostory
Authors
Chunglae Cho, Seungjae Shin, Hongseok Jeon, Seunghyun Yoon
Issue Date
2021-12
Citation
Global Communications Conference (GLOBECOM) 2021, pp.1-6
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1109/GLOBECOM46510.2021.9685845
Abstract
Utility-driven caching opened up a new design opportunity for caching algorithms by modeling the admission and eviction control as a utility maximization process with essential support for service differentiation. Nevertheless, there is still to go in terms of adaptability to changing environment. Slow convergence to an optimal state may degrade actual user-experienced utility, which gets even worse in non-stationary scenarios where cache control should be adaptive to time-varying content request traffic. This paper proposes to exploit deep reinforcement learning (DRL) to enhance the adaptability of utility-driven time-to-live (TTL)-based caching. Employing DRL with long short-term memory helps a caching agent learn how it adapts to the temporal correlation of content popularities to shorten the transient-state before the optimal steady-state. In addition, we elaborately design the state and action spaces of DRL to overcome the curse of dimensionality, which is one of the most frequently raised issues in machine learning-based approaches. Experimental results show that policies trained by DRL can outperform the conventional utility-driven caching algorithm under some non-stationary environments where content request traffic changes rapidly.
KSP Keywords
Changing environment, Deep reinforcement learning, Learning-based, Long-short term memory(LSTM), Nonstationary Environment, Optimal state, Reinforcement Learning(RL), Service differentiation, Slow convergence, Temporal Correlation, Time-to-live(TTL)