Optimization-driven Hierarchical Deep Reinforcement Learning for Hybrid Relaying Communications

Publisher:
IEEE
Publication Type:
Conference Proceeding
Citation:
2020 IEEE Wireless Communications and Networking Conference (WCNC), 2020, 2020-May, pp. 1-6
Issue Date:
2020-06-19
Filename Description Size
09120470.pdfPublished version219.56 kB
Adobe PDF
Full metadata record
In this paper, we employ multiple wireless-powered user devices as wireless relays to assist information transmission from a multi-antenna access point to a single-antenna receiver. To improve energy efficiency, we design a hybrid relaying communication strategy in which wireless relays are allowed to operate in either the passive mode via backscatter communications or the active mode via RF communications, depending on their channel conditions and energy states. We aim to maximize the overall SNR by jointly optimizing the access point's beamforming strategy as well as individual relays' radio modes and operating parameters. Due to the non-convex and combinatorial structure of the SNR maximization problem, we develop a deep reinforcement learning approach that adapts the beamforming and relaying strategies dynamically. In particular, we propose a novel optimization-driven hierarchical deep deterministic policy gradient (H-DDPG) approach that integrates the model-based optimization into the framework of conventional DDPG approach. It decomposes the discrete relay mode selection into the outer-loop by using deep Q-network (DQN) algorithm and then optimizes the continuous beamforming and relays' operating parameters by using the inner-loop DDPG algorithm. Simulation results reveal that the H-DDPG is robust to the hyper parameters and can speed up the learning process compared to the conventional DDPG approach.
Please use this identifier to cite or link to this item: