Transcription of Infinite-Horizon Discounted Markov Decision …
{{id}} {{{paragraph}}}
Infinite-Horizon Discounted Markov Decision processes Dan Zhang Leeds School of Business University of Colorado at Boulder Dan Zhang, Spring 2012 Infinite Horizon Discounted MDP 1. Outline The expected total Discounted reward Policy evaluation Optimality equations Value iteration Policy iteration Linear Programming Dan Zhang, Spring 2012 Infinite Horizon Discounted MDP 2. Expected Total Reward Criterion Let = (d1 , d2 , .. ) HR. Starting at a state s, using policy leads to a sequence of state-action pairs {Xt , Yt }. The sequence of rewards is given by {Rt rt (Xt , Yt ) : t = 1, 2, .. }. Let [0, 1) be the discount factor The expected total rewards from policy starting in state s is given by " N #. X. v (s) lim E s t 1 r (Xt , Yt ) . N . t=1. The limit above exists when r ( ) is bounded; , sups S,a As |r (s, a)| = M <.]
In nite-Horizon Discounted Markov Decision Processes Dan Zhang Leeds School of Business University of Colorado at Boulder Dan Zhang, Spring 2012 In nite Horizon Discounted MDP 1
Domain:
Source:
Link to this page:
Please notify us if you found a problem with this document:
{{id}} {{{paragraph}}}
Analysis of discounted cash flow (DCF), Analysis of discounted cash flow (DCF) approach, Price List We Deliver Discounted pricing, Estimate the Long-Term Growth, Estimate the Long-Term Growth Rate in the Discounted, Discounted Cash Flow Methodology, Discounted, FLOW ANALYSIS FOR, Flow analysis for market valuations, Discounted Cash Flow Analysis, COMMONLY USED METHODS OF VALUATION