Optimization-driven Hierarchical Deep Reinforcement Learning for Hybrid Relaying Communications
- Publisher:
- IEEE
- Publication Type:
- Conference Proceeding
- Citation:
- 2020 IEEE Wireless Communications and Networking Conference (WCNC), 2020, 2020-May, pp. 1-6
- Issue Date:
- 2020-06-19
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
09120470.pdf | Published version | 219.56 kB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
In this paper, we employ multiple wireless-powered user devices as wireless relays to assist information transmission from a multi-antenna access point to a single-antenna receiver. To improve energy efficiency, we design a hybrid relaying communication strategy in which wireless relays are allowed to operate in either the passive mode via backscatter communications or the active mode via RF communications, depending on their channel conditions and energy states. We aim to maximize the overall SNR by jointly optimizing the access point's beamforming strategy as well as individual relays' radio modes and operating parameters. Due to the non-convex and combinatorial structure of the SNR maximization problem, we develop a deep reinforcement learning approach that adapts the beamforming and relaying strategies dynamically. In particular, we propose a novel optimization-driven hierarchical deep deterministic policy gradient (H-DDPG) approach that integrates the model-based optimization into the framework of conventional DDPG approach. It decomposes the discrete relay mode selection into the outer-loop by using deep Q-network (DQN) algorithm and then optimizes the continuous beamforming and relays' operating parameters by using the inner-loop DDPG algorithm. Simulation results reveal that the H-DDPG is robust to the hyper parameters and can speed up the learning process compared to the conventional DDPG approach.
Please use this identifier to cite or link to this item: