Convex Optimization
Resource allocation and cross-layer control in wireless networks
Foundations and Trends® in Networking
Stochastic Learning and Optimization: A Sensitivity-Based Approach (International Series on Discrete Event Dynamic Systems)
Approximate Dynamic Programming: Solving the Curses of Dimensionality (Wiley Series in Probability and Statistics)
Dynamic Programming and Optimal Control, Vol. II
Dynamic Programming and Optimal Control, Vol. II
Reinforcement Learning: A Tutorial Survey and Recent Advances
INFORMS Journal on Computing
Delay-optimal power and precoder adaptation for multi-stream MIMO systems
IEEE Transactions on Wireless Communications
Adaptive multi-pattern reuse in multi-cell networks
WiOPT'09 Proceedings of the 7th international conference on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks
IEEE Transactions on Signal Processing
Binary Power Control for Sum Rate Maximization over Multiple Interfering Links
IEEE Transactions on Wireless Communications
Optimal Power and Rate Control for Minimal Average Delay: The Single-User Case
IEEE Transactions on Information Theory
Distributed Downlink Beamforming With Cooperative Base Stations
IEEE Transactions on Information Theory
A tutorial on decomposition methods for network utility maximization
IEEE Journal on Selected Areas in Communications
Hi-index | 0.01 |
In this paper, we propose a distributive queue-aware intra-cell user scheduling and inter-cell interference (ICI) management control design for a delay-optimal celluar downlink system with M base stations (BSs), and K users in each cell. Each BS has K downlink queues for K users respectively with heterogeneous arrivals and delay requirements. The ICI management control is adaptive to joint queue state information (QSI) over a slow time scale, while the user scheduling control is adaptive to both the joint QSI and the joint channel state information (CSI) over a faster time scale. We show that the problem can be modeled as an infinite horizon average cost Partially Observed Markov Decision Problem (POMDP), which is NP-hard in general. By exploiting the special structure of the problem, we shall derive an equivalent Bellman equation to solve the POMDP problem. To address the distributive requirement and the issue of dimensionality and computation complexity, we derive a distributive online stochastic learning algorithm, which only requires local QSI and local CSI at each of the M BSs. We show that the proposed learning algorithm converges almost-surely (with probability 1) and has significant gain compared with various baselines. The proposed solution only has linear complexity order O(MK).