Searching for large state space : 2 results found | RSS Feed for this search

Description

This course covers the basic models and solution techniques for problems of sequential decision making under uncertainty (stochastic control). We will consider optimal control of a dynamical system over both a finite and an infinite number of stages (finite and infinite horizon). We will also discuss some approximation methods for problems involving large state spaces. Applications of dynamic programming in a variety of fields will be covered in recitations. This course covers the basic models and solution techniques for problems of sequential decision making under uncertainty (stochastic control). We will consider optimal control of a dynamical system over both a finite and an infinite number of stages (finite and infinite horizon). We will also discuss some approximation methods for problems involving large state spaces. Applications of dynamic programming in a variety of fields will be covered in recitations.Subjects

dynamic programming | dynamic programming | stochastic control | stochastic control | decision making | decision making | uncertainty | uncertainty | sequential decision making | sequential decision making | finite horizon | finite horizon | infinite horizon | infinite horizon | approximation methods | approximation methods | state space | state space | large state space | large state space | optimal control | optimal control | dynamical system | dynamical system | dynamic programming and optimal control | dynamic programming and optimal control | deterministic systems | deterministic systems | shortest path | shortest path | state information | state information | rollout | rollout | stochastic shortest path | stochastic shortest path | approximate dynamic programming | approximate dynamic programmingLicense

Content within individual OCW courses is (c) by the individual authors unless otherwise noted. MIT OpenCourseWare materials are licensed by the Massachusetts Institute of Technology under a Creative Commons License (Attribution-NonCommercial-ShareAlike). For further information see http://ocw.mit.edu/terms/index.htmSite sourced from

http://ocw.mit.edu/rss/all/mit-allarchivedcourses.xmlAttribution

Click to get HTML | Click to get attribution | Click to get URLAll metadata

See all metadata6.231 Dynamic Programming and Stochastic Control (MIT)

Description

This course covers the basic models and solution techniques for problems of sequential decision making under uncertainty (stochastic control). We will consider optimal control of a dynamical system over both a finite and an infinite number of stages (finite and infinite horizon). We will also discuss some approximation methods for problems involving large state spaces. Applications of dynamic programming in a variety of fields will be covered in recitations.Subjects

dynamic programming | stochastic control | decision making | uncertainty | sequential decision making | finite horizon | infinite horizon | approximation methods | state space | large state space | optimal control | dynamical system | dynamic programming and optimal control | deterministic systems | shortest path | state information | rollout | stochastic shortest path | approximate dynamic programmingLicense

Content within individual OCW courses is (c) by the individual authors unless otherwise noted. MIT OpenCourseWare materials are licensed by the Massachusetts Institute of Technology under a Creative Commons License (Attribution-NonCommercial-ShareAlike). For further information see https://ocw.mit.edu/terms/index.htmSite sourced from

https://ocw.mit.edu/rss/all/mit-allarchivedcourses.xmlAttribution

Click to get HTML | Click to get attribution | Click to get URLAll metadata

See all metadata