A Tutorial on Linear Function Approximators for Dynamic Programming and Reinforcement Learning (Paperback)

, , , , ,
A Markov Decision Process (MDP) is a natural framework for formulating sequential decision-making problems under uncertainty. In recent years, researchers have greatly advanced algorithms for learning and acting in MDPs. This book reviews such algorithms, beginning with well-known dynamic programming methods for solving MDPs such as policy iteration and value iteration, then describes approximate dynamic programming methods such as trajectory based value iteration, and finally moves to reinforcement learning methods such as Q-Learning, SARSA, and least-squares policy iteration. It describes algorithms in a unified framework, giving pseudocode together with memory and iteration complexity analysis for each. Empirical evaluations of these techniques, with four representations across four domains, provide insight into how these algorithms perform with various feature sets in terms of running time and performance. This tutorial provides practical guidance for researchers seeking to extend DP and RL techniques to larger domains through linear value function approximation. The practical algorithms and empirical successes outlined also form a guide for practitioners trying to weigh computational costs, accuracy requirements, and representational concerns. Decision making in large domains will always be challenging, but with the tools presented here this challenge is not insurmountable.

R1,701

Or split into 4x interest-free payments of 25% on orders over R50
Learn more

Discovery Miles17010
Mobicred@R159pm x 12* Mobicred Info
Free Delivery
Delivery AdviceShips in 10 - 15 working days



Product Description

A Markov Decision Process (MDP) is a natural framework for formulating sequential decision-making problems under uncertainty. In recent years, researchers have greatly advanced algorithms for learning and acting in MDPs. This book reviews such algorithms, beginning with well-known dynamic programming methods for solving MDPs such as policy iteration and value iteration, then describes approximate dynamic programming methods such as trajectory based value iteration, and finally moves to reinforcement learning methods such as Q-Learning, SARSA, and least-squares policy iteration. It describes algorithms in a unified framework, giving pseudocode together with memory and iteration complexity analysis for each. Empirical evaluations of these techniques, with four representations across four domains, provide insight into how these algorithms perform with various feature sets in terms of running time and performance. This tutorial provides practical guidance for researchers seeking to extend DP and RL techniques to larger domains through linear value function approximation. The practical algorithms and empirical successes outlined also form a guide for practitioners trying to weigh computational costs, accuracy requirements, and representational concerns. Decision making in large domains will always be challenging, but with the tools presented here this challenge is not insurmountable.

Customer Reviews

No reviews or ratings yet - be the first to create one!

Product Details

General

Imprint

Now Publishers Inc

Country of origin

United States

Series

Foundations and Trends (R) in Machine Learning

Release date

December 2013

Availability

Expected to ship within 10 - 15 working days

First published

2013

Authors

, , , , ,

Dimensions

234 x 156 x 5mm (L x W x T)

Format

Paperback

Pages

92

ISBN-13

978-1-60198-760-0

Barcode

9781601987600

Categories

LSN

1-60198-760-9



Trending On Loot