Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
LQ-Routing Protocol for Mobile Ad-Hoc Networks
Proceedings of the Fourth Annual ACIS International Conference on Computer and Information Science
Detecting Primary Signals for Efficient Utilization of Spectrum Using Q-Learning
ITNG '08 Proceedings of the Fifth International Conference on Information Technology: New Generations
CRAHNs: Cognitive radio ad hoc networks
Ad Hoc Networks
Reinforcement learning: a survey
Journal of Artificial Intelligence Research
Spectrum management of cognitive radio using multi-agent reinforcement learning
Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: Industry track
Applications of Machine Learning to Cognitive Radio Networks
IEEE Wireless Communications
A Comprehensive Survey of Multiagent Reinforcement Learning
IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews
Using feedback in collaborative reinforcement learning to adaptively optimize MANET routing
IEEE Transactions on Systems, Man, and Cybernetics, Part A: Systems and Humans
A dynamic channel assignment policy through Q-learning
IEEE Transactions on Neural Networks
Hi-index | 0.00 |
Cognitive Radio Ad Hoc Networks (CRAHNs) must identify the best operational characteristics based on the local spectrum availability, reachability with other nodes, choice of spectrum, while maintaining an acceptable end-to-end performance. The distributed nature of the operation forces each node to act autonomously, and yet has a goal of optimizing the overall network performance. These unique characteristics of CRAHNs make reinforcement learning (RL) techniques an attractive choice as a tool for protocol design. In this paper, we survey the state-of-the-art in the existing RL schemes that can be applied to CRAHNs, and propose modifications from the viewpoint of routing, and link layer spectrum-aware operations. We provide a framework of applying RL techniques for joint power and spectrum allocation as an example of Q-learning. Finally, through simulation study, we demonstrate the benefits of using RL schemes in dynamic spectrum conditions.