Back to Search Start Over

Q -Learning Algorithms for Constrained Markov Decision Processes With Randomized Monotone Policies: Application to MIMO Transmission Control.

Authors :
Djonin, Dejan V.
Krishnamurthy, Vikram
Source :
IEEE Transactions on Signal Processing. May2007 Part 2, Vol. 55 Issue 5, p2170-2181. 12p. 2 Diagrams, 9 Graphs.
Publication Year :
2007

Abstract

This paper presents novel Q-learning based stochastic control algorithms for rate and power control in V-BLAST transmission systems. The algorithms exploit the supermodularity and monotonic structure results derived in the companion paper. Rate and power control problem is posed as a stochastic optimization problem with the goal of minimizing the average transmission power under the constraint on the average delay that can be interpreted as the quality of service requirement of a given application. Standard Q-learning algorithm is modified to handle the constraints so that it can adaptively learn structured optimal policy for unknown channel/traffic statistics. We discuss the convergence of the proposed algorithms and explore their properties in simulations. To address the issue of unknown transmission costs in an unknown time-varying environment, we propose the variant of Q-learning algorithm in which power costs are estimated in online fashion, and we show that this algorithm converges to the optimal solution as long as the power cost estimates are asymptotically unbiased [ABSTRACT FROM PUBLISHER]

Details

Language :
English
ISSN :
1053587X
Volume :
55
Issue :
5
Database :
Academic Search Index
Journal :
IEEE Transactions on Signal Processing
Publication Type :
Academic Journal
Accession number :
52037317
Full Text :
https://doi.org/10.1109/TSP.2007.893228