A policy gradient method for semi-Markov decision processes with application to call admission control
From MaRDI portal
Publication:859693
DOI10.1016/J.EJOR.2006.02.023zbMath1163.90790OpenAlexW2169293926MaRDI QIDQ859693
Sumeetpal S. Singh, Arnaud Doucet, Vladislav B. Tadić
Publication date: 16 January 2007
Published in: European Journal of Operational Research (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.ejor.2006.02.023
Related Items (9)
A reinforcement-learning approach for admission control in distributed network service systems ⋮ A minimization problem of the risk probability in first passage semi-Markov decision processes with loss rates ⋮ Finite horizon semi-Markov decision processes with application to maintenance systems ⋮ Semiconductor final test scheduling with Sarsa\((\lambda , k)\) algorithm ⋮ Approximate dynamic programming for capacity allocation in the service industry ⋮ Performance analysis for controlled semi-Markov systems with application to maintenance ⋮ Minimizing mean weighted tardiness in unrelated parallel machine scheduling with reinforcement learning ⋮ Look-ahead control of conveyor-serviced production station by using potential-based online policy iteration ⋮ FLOW SHOP SCHEDULING WITH REINFORCEMENT LEARNING
Cites Work
- Markov chains and stochastic stability
- Reinforcement learning for long-run average cost.
- Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning
- Gradient Convergence in Gradient methods with Errors
- Semi-markov decision problems and performance sensitivity analysis
- Integrated voice/data call admission control for wireless DS-CDMA systems
- On the convergence of temporal-difference learning with linear function approximation
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
This page was built for publication: A policy gradient method for semi-Markov decision processes with application to call admission control