Deep Reinforcement Learning for Reconfigurable Intelligent Surfaces and UAV Empowered Smart 6G Communications
English
By (author): Antonino Masaracchia Khoi Khac Nguyen Trung Q. Duong Vishal Sharma
Reconfigurable intelligent surface (RIS) has emerged as a cutting-edge technology for beyond 5G and 6G networks due to its low-cost hardware production, nearly passive nature, easy deployment, communication without new waves, and energy-saving benefits. Unmanned aerial vehicle (UAV)-assisted wireless networks significantly enhance network coverage.
Resource allocation and real-time decision-making optimisation play a pivotal role in approaching the optimal performance in UAV- and RIS-aided wireless communications. But the existing contributions typically assume having a static environment and often ignore the stringent flight time constraints in real-life applications. It is crucial to improve the decision-making time for meeting the stringent requirements of UAV-assisted wireless networks. Deep reinforcement learning (DRL), which is a combination of reinforcement learning and neural networks, is used to maximise network performance, reduce power consumption, and improve the processing time for real-time. DRL algorithms can help UAVs and RIS work fully autonomously, reduce energy consumption and operate optimally in an unexpected environment.
This co-authored book explores the many challenges arising from real-time and autonomous decision-making for 6G. The goal is to provide readers with comprehensive insights into the models and techniques of deep reinforcement learning and its applications in 6G networks and internet-of-things with the support of UAVs and RIS.
Deep Reinforcement Learning for Reconfigurable Intelligent Surfaces and UAV Empowered Smart 6G Communications is aimed at a wide audience of researchers, practitioners, scientists, professors and advanced students in engineering, computer science, information technology, and communication engineering, and networking and ubiquitous computing professionals.
See moreWill deliver when available. Publication date 01 Dec 2024