A new reward model for MDP state aggregation with application to CAC and Routing

An optimal solution of the call admission control and routing problem in multi‐service loss networks, in terms of average reward per time unit, is possible by modeling the network behavior as a Markov decision process (MDP). However, even after applying the standard link independence assumption, the...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:European transactions on telecommunications 2005-11, Vol.16 (6), p.495-508
Hauptverfasser: Nordström, Ernst, Carlström, Jakob
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:An optimal solution of the call admission control and routing problem in multi‐service loss networks, in terms of average reward per time unit, is possible by modeling the network behavior as a Markov decision process (MDP). However, even after applying the standard link independence assumption, the solution of the corresponding set of link problems may involve considerable numerical computation. In this paper, we study an approximate MDP framework on the link level, where vector‐valued MDP states are mapped into a set of aggregate scalar MDP states corresponding to link occupancies. In particular, we propose a new model of the expected reward for admitting a call on the network. Compared to Krishnan's and Hübner's method 11, our reward model more accurately reflects the bandwidth occupancy by different call categories. The exact and approximate link MDP frameworks are compared by simulations, and the results show that the proposed link reward model significantly improves the performance of Krishnan's and Hübner's method. Copyright © 2004 AEI.
ISSN:1124-318X
1541-8251
DOI:10.1002/ett.1007