UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Energy Minimization in D2D-Assisted Cache-Enabled Internet of Things: A Deep Reinforcement Learning Approach

Tang, J; Tang, H; Zhang, X; Cumanan, K; Chen, G; Wong, KK; Chambers, JA; (2020) Energy Minimization in D2D-Assisted Cache-Enabled Internet of Things: A Deep Reinforcement Learning Approach. IEEE Transactions on Industrial Informatics , 16 (8) pp. 5412-5423. 10.1109/TII.2019.2954127. Green open access

[thumbnail of ALL_TII_19_3001.R1.pdf]
Preview
Text
ALL_TII_19_3001.R1.pdf - Accepted Version

Download (701kB) | Preview

Abstract

Mobile edge caching (MEC) and device-to-device (D2D) communications are two potential technologies to resolve traffic overload problems in the Internet of Things. Previous works usually investigate them separately with MEC for traffic offloading and D2D for information transmission. In this article, a joint framework consisting of MEC and cache-enabled D2D communications is proposed to minimize the energy cost of systematic traffic transmission, where file popularity and user preference are the critical criteria for small base stations (SBSs) and user devices, respectively. Under this framework, we propose a novel caching strategy, where the Markov decision process is applied to model the requesting behaviors. A novel scheme based on reinforcement learning (RL) is proposed to reveal the popularity of files as well as users' preference. In particular, a Q-learning algorithm and a deep Q-network algorithm are, respectively, applied to user devices and the SBS due to different complexities of status. To save the energy cost of systematic traffic transmission, users acquire partial traffic through D2D communications based on the cached contents and user distribution. Taking the memory limits, D2D available files, and status changing into consideration, the proposed RL algorithm enables user devices and the SBS to prefetch the optimal files while learning, which can reduce the energy cost significantly. Simulation results demonstrate the superior energy saving performance of the proposed RL-based algorithm over other existing methods under various conditions.

Type: Article
Title: Energy Minimization in D2D-Assisted Cache-Enabled Internet of Things: A Deep Reinforcement Learning Approach
Open access status: An open access version is available from UCL Discovery
DOI: 10.1109/TII.2019.2954127
Publisher version: https://doi.org/10.1109/TII.2019.2954127
Language: English
Additional information: This version is the author accepted manuscript. For information on re-use, please refer to the publisher’s terms and conditions.
Keywords: Content caching, deep $Q$ -network (DQN), device-to-device (D2D) communications, Internet of Things (IoT), $Q$ -learning (QL)
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL BEAMS
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Electronic and Electrical Eng
URI: https://discovery.ucl.ac.uk/id/eprint/10098303
Downloads since deposit
247Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item