TY - JOUR
T1 - Green Deep Reinforcement Learning for Radio Resource Management
T2 - Architecture, Algorithm Compression, and Challenges
AU - Du, Zhiyong
AU - Deng, Yansha
AU - Guo, Weisi
AU - Nallanathan, Arumugam
AU - Wu, Qihui
N1 - Funding Information:
This article is partly funded by EC H2020 grant 778305 (Data Aware Wireless Networks for IoE) and grant 892221 (Green Machine Learning for 5G) and National Natural Science Foundation of China grant 61601490. The corresponding author of this article is Weisi Guo.
Publisher Copyright:
© 2005-2012 IEEE.
Copyright:
Copyright 2021 Elsevier B.V., All rights reserved.
PY - 2021/3
Y1 - 2021/3
N2 - Artificial intelligence (AI) heralds a step-change in wireless networks but may also cause irreversible environmental damage due to its high energy consumption. Here, we address this challenge in the context of 5G and beyond, where there is a complexity explosion in radio resource management (RRM). For high-dimensional RRM problems in a dynamic environment, deep reinforcement learning (DRL) provides a powerful tool for scalable optimization, but it consumes a large amount of energy over time and risks compromising progress made in green radio research. This article reviews and analyzes how to achieve green DRL for RRM via both architecture and algorithm innovations. Architecturally, a cloudbased training and distributed decision-making DRL scheme is proposed, where RRM entities can make lightweight, deep, local decisions while being assisted by on-cloud training and updating. At the algorithm level, compression approaches are introduced for both deep neural networks (DNNs) and the underlying Markov decision processes (MDPs), enabling accurate lowdimensional representations of challenges. To scale learning across geographic areas, a spatial transfer learning scheme is proposed to further promote the learning efficiency of distributed DRL entities by exploiting the traffic demand correlations. Together, our proposed architecture and algorithms provide a vision for green and on-demand DRL capability.
AB - Artificial intelligence (AI) heralds a step-change in wireless networks but may also cause irreversible environmental damage due to its high energy consumption. Here, we address this challenge in the context of 5G and beyond, where there is a complexity explosion in radio resource management (RRM). For high-dimensional RRM problems in a dynamic environment, deep reinforcement learning (DRL) provides a powerful tool for scalable optimization, but it consumes a large amount of energy over time and risks compromising progress made in green radio research. This article reviews and analyzes how to achieve green DRL for RRM via both architecture and algorithm innovations. Architecturally, a cloudbased training and distributed decision-making DRL scheme is proposed, where RRM entities can make lightweight, deep, local decisions while being assisted by on-cloud training and updating. At the algorithm level, compression approaches are introduced for both deep neural networks (DNNs) and the underlying Markov decision processes (MDPs), enabling accurate lowdimensional representations of challenges. To scale learning across geographic areas, a spatial transfer learning scheme is proposed to further promote the learning efficiency of distributed DRL entities by exploiting the traffic demand correlations. Together, our proposed architecture and algorithms provide a vision for green and on-demand DRL capability.
KW - Training
KW - Optimization
KW - Green products
KW - Energy consumption
KW - Vehicle dynamics
KW - Machine learning
KW - Wireless networks
UR - http://www.scopus.com/inward/record.url?scp=85101470713&partnerID=8YFLogxK
U2 - 10.1109/MVT.2020.3015184
DO - 10.1109/MVT.2020.3015184
M3 - Article
AN - SCOPUS:85101470713
SN - 1556-6072
VL - 16
SP - 29
EP - 39
JO - IEEE Vehicular Technology Magazine
JF - IEEE Vehicular Technology Magazine
IS - 1
M1 - 9205233
ER -