DSpace Repository

The optimal control of Markovian processes

Show simple item record

dc.contributor.author Miller, Roy A
dc.date.accessioned 2011-06-21T01:54:18Z
dc.date.accessioned 2022-10-26T20:49:13Z
dc.date.available 2011-06-21T01:54:18Z
dc.date.available 2022-10-26T20:49:13Z
dc.date.copyright 1977
dc.date.issued 1977
dc.identifier.uri https://ir.wgtn.ac.nz/handle/123456789/24882
dc.description.abstract With the development of the theory of dynamic programming, considerable interest was generated in Markovian decision processes and since 1960 several research papers have been published on this topic. Even if solutions were difficult to obtain via the dynamic programming method, it was found that certain problems arising from the fields of systems engineering, operations research and optimal control could be formulated as dynamic programming problems. This thesis gives an expository account of the mathematical theory and algorithms involved in the optimal control of Markovian dynamic systems. Such a system is observed periodically and classified into one of a possible number of states. After each classification one of a possible number of actions is taken. The sequence of implemented actions (called a policy) interacts in a probabilistic manner to effect the evolution of the system. The mathematical abstraction of the process is called a Markovian decision process. Throughout the thesis optimization is considered relative to two objective functions - the total expected discounted return and the average return per unit time. Emphasis is placed on the existence of optimal policies and the methods of determining such policies. Some illustrative examples are included. Chapter I begins with a definition of the discrete time problem over a finite planning horizon. From then on the problem is considered over an infinite planning horizon. The state and action spaces are finite. A policy iterative technique called the Howard policy improvement routine is described and it is shown that these problems can be formulated and solved as linear programming problems. In Chapter II we consider the discrete time problem when the state space is a denumerable set. There it is pointed out that under the average return criterion an optimal policy may not exist. Sufficient conditions for the existence of optimal policies are given and use is made of the principle of contraction mapping in obtaining solutions. In Chapter III we discuss semi-Markovian decision processes. The continuous time Markovian decision process is considered as a special case of the semi-Markovian process. It is shown that the policy iterative technique can be applied to the continuous problem. It is also shown how the continuous problem may be formulated into a linear programming problem by first transforming the continuous problem into a discrete one and formulating the resulting discrete problem into a linear programming one. Finally, in Chapter IV we consider the continuous time problem under a denumerable state space. We discuss results analogous to those for the discrete case of Chapter II. en_NZ
dc.format pdf en_NZ
dc.language en_NZ
dc.language.iso en_NZ
dc.publisher Te Herenga Waka—Victoria University of Wellington en_NZ
dc.title The optimal control of Markovian processes en_NZ
dc.type Text en_NZ
vuwschema.type.vuw Awarded Research Masters Thesis en_NZ
thesis.degree.discipline Mathematics en_NZ
thesis.degree.grantor Te Herenga Waka—Victoria University of Wellington en_NZ
thesis.degree.level Masters en_NZ
thesis.degree.name Master of Science en_NZ


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Browse

My Account