A dynamic clustering technique based on deep reinforcement learning for Internet of vehicles

A Sharif, JP Li, MA Saleem, G Manogran… - Journal of Intelligent …, 2021 - Springer
A Sharif, JP Li, MA Saleem, G Manogran, S Kadry, A Basit, MA Khan
Journal of Intelligent Manufacturing, 2021Springer
Abstract The Internet of Vehicles (IoV) is a communication paradigm that connects the
vehicles to the Internet for transferring information between the networks. One of the key
challenges in IoV is the management of a massive amount of traffic generated from a large
number of connected IoT-based vehicles. Network clustering strategies have been proposed
to solve the challenges of traffic management in IoV networks. Traditional optimization
approaches have been proposed to manage the resources of the network efficiently …
Abstract
The Internet of Vehicles (IoV) is a communication paradigm that connects the vehicles to the Internet for transferring information between the networks. One of the key challenges in IoV is the management of a massive amount of traffic generated from a large number of connected IoT-based vehicles. Network clustering strategies have been proposed to solve the challenges of traffic management in IoV networks. Traditional optimization approaches have been proposed to manage the resources of the network efficiently. However, the nature of next-generation IoV environment is highly dynamic, and the existing optimization technique cannot precisely formulate the dynamic characteristic of IoV networks. Reinforcement learning is a model-free technique where an agent learns from its environment for learning the optimal policies. We propose an experience-driven approach based on an Actor-Critic based Deep Reinforcement learning framework (AC-DRL) for efficiently selecting the cluster head (CH) for managing the resources of the network considering the noisy nature of IoV environment. The agent in the proposed AC-DRL can efficiently approximate and learn the state-action value function of the actor and action function of the critic for selecting the CH considering the dynamic condition of the network.The experimental results show an improvement of 28% and 15% respectively, in terms of satisfying the SLA requirement and 35% and 14% improvement in throughput compared to the static and DQN approaches.
Springer
以上显示的是最相近的搜索结果。 查看全部搜索结果