Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Lightweight Reinforcement Learning for Priority-Aware Spectrum Management in Vehicular IoT Networks
by
Nauman, Ali
, Iqbal, Adeel
, Khurshaid, Tahir
in
Communication
/ Decision making
/ Design
/ Energy consumption
/ Energy efficiency
/ Heuristic
/ Infrastructure
/ Internet of Things
/ IoT
/ Markov Decision Process
/ Markov processes
/ Privacy
/ QoS
/ Real time
/ reinforcement learning
/ Scheduling
/ Traffic
/ V-IoT
/ Vehicles
2025
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Lightweight Reinforcement Learning for Priority-Aware Spectrum Management in Vehicular IoT Networks
by
Nauman, Ali
, Iqbal, Adeel
, Khurshaid, Tahir
in
Communication
/ Decision making
/ Design
/ Energy consumption
/ Energy efficiency
/ Heuristic
/ Infrastructure
/ Internet of Things
/ IoT
/ Markov Decision Process
/ Markov processes
/ Privacy
/ QoS
/ Real time
/ reinforcement learning
/ Scheduling
/ Traffic
/ V-IoT
/ Vehicles
2025
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Lightweight Reinforcement Learning for Priority-Aware Spectrum Management in Vehicular IoT Networks
by
Nauman, Ali
, Iqbal, Adeel
, Khurshaid, Tahir
in
Communication
/ Decision making
/ Design
/ Energy consumption
/ Energy efficiency
/ Heuristic
/ Infrastructure
/ Internet of Things
/ IoT
/ Markov Decision Process
/ Markov processes
/ Privacy
/ QoS
/ Real time
/ reinforcement learning
/ Scheduling
/ Traffic
/ V-IoT
/ Vehicles
2025
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Lightweight Reinforcement Learning for Priority-Aware Spectrum Management in Vehicular IoT Networks
Journal Article
Lightweight Reinforcement Learning for Priority-Aware Spectrum Management in Vehicular IoT Networks
2025
Request Book From Autostore
and Choose the Collection Method
Overview
The Vehicular Internet of Things (V-IoT) has emerged as a cornerstone of next-generation intelligent transportation systems (ITSs), enabling applications ranging from safety-critical collision avoidance and cooperative awareness to infotainment and fleet management. These heterogeneous services impose stringent quality-of-service (QoS) demands for latency, reliability, and fairness while competing for limited and dynamically varying spectrum resources. Conventional schedulers, such as round-robin or static priority queues, lack adaptability, whereas deep reinforcement learning (DRL) solutions, though powerful, remain computationally intensive and unsuitable for real-time roadside unit (RSU) deployment. This paper proposes a lightweight and interpretable reinforcement learning (RL)-based spectrum management framework for Vehicular Internet of Things (V-IoT) networks. Two enhanced Q-Learning variants are introduced: a Value-Prioritized Action Double Q-Learning with Constraints (VPADQ-C) algorithm that enforces reliability and blocking constraints through a Constrained Markov Decision Process (CMDP) with online primal–dual optimization, and a contextual Q-Learning with Upper Confidence Bound (Q-UCB) method that integrates uncertainty-aware exploration and a Success-Rate Prior (SRP) to accelerate convergence. A Risk-Aware Heuristic baseline is also designed as a transparent, low-complexity benchmark to illustrate the interpretability–performance trade-off between rule-based and learning-driven approaches. A comprehensive simulation framework incorporating heterogeneous traffic classes, physical-layer fading, and energy-consumption dynamics is developed to evaluate throughput, delay, blocking probability, fairness, and energy efficiency. The results demonstrate that the proposed methods consistently outperform conventional Q-Learning and Double Q-Learning methods. VPADQ-C achieves the highest energy efficiency (≈8.425×107 bits/J) and reduces interruption probability by over 60%, while Q-UCB achieves the fastest convergence (within ≈190 episodes), lowest blocking probability (≈0.0135), and lowest mean delay (≈0.351 ms). Both schemes maintain fairness near 0.364, preserve throughput around 28 Mbps, and exhibit sublinear training-time scaling with O(1) per-update complexity and O(N2) overall runtime growth. Scalability analysis confirms that the proposed frameworks sustain URLLC-grade latency (<0.2 ms) and reliability under dense vehicular loads, validating their suitability for real-time, large-scale V-IoT deployments.
Publisher
MDPI AG
Subject
This website uses cookies to ensure you get the best experience on our website.