Documents
Presentation Slides
Cooperative Scenarios For Multi-agent Reinforcement learning In Wireless Edge Caching
- Citation Author(s):
- Submitted by:
- Navneet Garg
- Last updated:
- 16 June 2021 - 4:16am
- Document Type:
- Presentation Slides
- Document Year:
- 2021
- Event:
- Presenters:
- Navneet Garg
- Paper Code:
- 2733
- Categories:
- Log in to post comments
Wireless edge caching is an important strategy to fulfill the demands in the next generation wireless systems. Recent studies have indicated that among a network of small base stations (SBSs), joint content placement improves the cache hit performance via reinforcement learning, since content requests are correlated across SBSs and files. In this paper, we investigate multi-agent reinforcement learning (MARL), and identify four scenarios for cooperation. These scenarios include full cooperation (S1), episodic cooperation (S2), distributed cooperation (S3), and independent operation (no-cooperation). MARL algorithms have been presented for each scenario. Simulations results for averaged normalized cache hits show that cooperation with one neighbor (S3) can improve the performance significantly closer to full-cooperation (S1). Scenario 2 shows the importance of frequent cooperation, when the level of cooperation is high, which depends on the number of SBSs.