Skip to search formSkip to main content
You are currently offline. Some features of the site may not work correctly.

Markov decision process

Known as: Value iteration, Policy iteration, Markov decision problems 
Markov decision processes (MDPs) provide a mathematical framework for modeling decision making in situations where outcomes are partly random and… Expand
Wikipedia

Papers overview

Semantic Scholar uses AI to extract papers important to this topic.
Review
2019
Review
2019
  • D. Bertsekas
  • IEEE/CAA Journal of Automatica Sinica
  • 2019
  • Corpus ID: 4881243
In this paper we discuss policy iteration methods for approximate solution of a finite-state discounted Markov decision problem… Expand
  • figure 1
  • figure 2
  • figure 3
  • figure 4
  • figure 5
Is this relevant?
Review
2017
Review
2017
Online learning is an important property of adaptive dynamic programming (ADP). Online observations contain plentiful dynamics… Expand
  • figure 1
  • table 1
  • figure 2
  • figure 3
  • figure 4
Is this relevant?
Review
2016
Review
2016
9 5 5 partially observable decision processes 9 5. partially observable total cost markov decision processes. partially observed… Expand
  • figure 1
  • figure 15.1
Is this relevant?
Highly Cited
2005
Highly Cited
2005
Optimal solutions to Markov decision problems may be very sensitive with respect to the state transition probabilities. In many… Expand
  • figure 1
  • figure 2
  • figure 3
  • figure 4
  • figure 5
Is this relevant?
Highly Cited
2004
Highly Cited
2004
A critical issue for the application of Markov decision processes (MDPs) to realistic problems is how the complexity of planning… Expand
  • figure 1
  • figure 2
Is this relevant?
Highly Cited
2001
Highly Cited
2001
  • F. Jensen
  • Statistics for Engineering and Information…
  • 2001
  • Corpus ID: 42979791
Probabilistic graphical models and decision graphs are powerful modeling tools for reasoning and decision making under… Expand
  • figure 1.1
  • figure 1.2
  • figure 1.3
Is this relevant?
Highly Cited
1999
Highly Cited
1999
INTRODUCTION Examples of Constrained Dynamic Control Problems On Solution Approaches for CMDPs with Expected Costs Other Types of… Expand
  • figure 8.1
  • figure 8.2
  • figure 8.3
Is this relevant?
Highly Cited
1996
Highly Cited
1996
1 Introduction.- 1.0 Background.- 1.1 Raison d'Etre and Limitations.- 1.2 A Menu of Courses and Prerequisites.- 1.3 For the… Expand
Is this relevant?
Highly Cited
1994
Highly Cited
1994
  • M. Puterman
  • Wiley Series in Probability and Statistics
  • 1994
  • Corpus ID: 122678161
From the Publisher: The past decade has seen considerable theoretical and applied research on Markov decision processes, as well… Expand
Is this relevant?
Highly Cited
1987
Highly Cited
1987
We investigate the complexity of the classical problem of optimal policy computation in Markov decision processes. All three… Expand
Is this relevant?