Summary
Approximate Dynamic Programming is a means of synthesizing nearoptimal policies for large scale stochastic control problems. We examine here the LP approach to approximate Dynamic Programming [98] which requires the solution of a linear program with a tractable number of variables but a potentially large number of constraints. Randomized constraint sampling is one means of dealing with such a program and results from [99] suggest that in fact, such a scheme is capable of producing good solutions to the linear program that arises in the context of approximate Dynamic Programming. We present here a summary of those results, and a case study wherein the technique is used to produce a controller for the game of Tetris. The case study highlights several practical issues concerning the applicability of the constraint sampling approach. We also demonstrate a controller that matches - and in some ways outperforms - controllers produced by other state of the art techniques for large-scale stochastic control.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag London Limited
About this chapter
Cite this chapter
Farias, V.F., Van Roy, B. (2006). Tetris: A Study of Randomized Constraint Sampling. In: Calafiore, G., Dabbene, F. (eds) Probabilistic and Randomized Methods for Design under Uncertainty. Springer, London. https://doi.org/10.1007/1-84628-095-8_6
Download citation
DOI: https://doi.org/10.1007/1-84628-095-8_6
Publisher Name: Springer, London
Print ISBN: 978-1-84628-094-8
Online ISBN: 978-1-84628-095-5
eBook Packages: EngineeringEngineering (R0)