Multi-Agent Reinforcement Learning for Cooperative Coded Caching via Homotopy Optimization
Creators
- 1. The Chinese University of Hong Kong
- 2. Nanjing University of Science and Technology
- 3. Royal Institute of Technology, KTH
- 4. Princeton University
Description
Introducing cooperative coded caching into small
cell networks is a promising approach to reducing traffic loads.
By encoding content via maximum distance separable (MDS)
codes, coded fragments can be collectively cached at small-cell
base stations (SBSs) to enhance caching efficiency. However, content
popularity is usually time-varying and unknown in practice.
As a result, cached content is anticipated to be intelligently
updated by taking into account limited caching storage and
interactive impacts among SBSs. In response to these challenges,
we propose a multi-agent deep reinforcement learning (DRL)
framework to intelligently update cached content in dynamic
environments. With the goal of minimizing long-term expected
fronthaul traffic loads, we first model dynamic coded caching as
a cooperative multi-agent Markov decision process. Owing to the
use of MDS coding, the resulting decision-making falls into a class
of constrained reinforcement learning problems with continuous
decision variables. To deal with this difficulty, we custom-build a
novel DRL algorithm by embedding homotopy optimization into
a deep deterministic policy gradient formalism. Next, to empower
the caching framework with an effective trade-off between complexity
and performance, we propose centralized, and partially
and fully decentralized caching controls by applying the derived
DRL approach. Simulation results demonstrate the superior
performance of the proposed multi-agent framework.
Files
Multi-Agent_Reinforcement_Learning_for_Cooperative_Coded_Caching_via_Homotopy_Optimization.pdf
Files
(2.0 MB)
Name | Size | Download all |
---|---|---|
md5:b973b606714d0c2468565a073b18c835
|
2.0 MB | Preview Download |