## Date of Award

August 2023

## Degree Type

Dissertation

## Degree Name

Doctor of Philosophy

## Department

Mathematics

## First Advisor

Richard H Stockbridge

## Keywords

control, probability, stochastic

## Abstract

Under consideration are convergence results between optimality criteria for two infinite-horizon stochastic control problems: the long-term average problem and the $\alpha$-discounted problem, where $\alpha \in (0,1]$ is a given discount rate. The objects under control are those stochastic processes that arise as (relaxed) solutions to a controlled martingale problem; and such controlled processes, subject to a given budget constraint, comprise the feasible sets for the two stochastic control problems.

In this dissertation, we define and characterize the expected occupation measures associated with each of these stochastic control problems, and then reformulate each problem as an equivalent linear program over a space of such measures. We then establish sufficient conditions under which the long-term average linear program can be ``asymptotically approximated'' by the $\alpha$-parameterized family of (suitably normalized) $\alpha$-discounted linear programs as $\alpha \downarrow 0$. This approach is what can be referred to as the vanishing discount method. To state these conditions precisely, our analysis turns to set-valued mappings called correspondences. In particular, once we establish the appropriate framework, we see that our main results can be stated in a manner similar to that of Berge's Theorem.

## Recommended Citation

Hospital, Brian, "The Vanishing Discount Method for Stochastic Control: A Linear Programming Approach" (2023). *Theses and Dissertations*. 3275.

https://dc.uwm.edu/etd/3275