## Abstract

In the literature, the problem of maximizing the expected discounted reward over all stopping rules has been explicitly solved for a number of reward functions (including (max{x, 0})^{v}, v > 0, in particular) when the underlying process is either a random walk in discrete time or a Levy process in continuous time. All of such reward functions are increasing and logconcave while the corresponding optimal stopping rules have the threshold form. In this paper, we explore the close connection between increasing and logconcave reward functions and optimal stopping rules of threshold form. In the discrete case, we show that if a reward function defined on ℤ is nonnegative, increasing and logconcave, then the optimal stopping rule is ofthreshold form provided the underlying random walk is skip-free to the right. In the continuous case, it is shown that for a reward function defined on ℝ which is nonnegative, increasing, logconcave and right-continuous, the optimal stopping rule is of threshold form provided the underlying process is a spectrally negative lévy process. Furthermore, we also establish the necessity of logconcavity and monotonicity of a reward function in order for the optimal stopping rule to be of threshold form in the discrete (continuous, resp.) case when the underlying process belongs to the class of Bernoulli random walks (Brownian motions, resp.) with a downward drift. These results together provide a partial characterization of the threshold structure of optimal stopping rules.

Original language | English |
---|---|

Article number | 120 |

Journal | Electronic Journal of Probability |

Volume | 19 |

DOIs | |

Publication status | Published - 2014 Dec 27 |

## All Science Journal Classification (ASJC) codes

- Statistics and Probability
- Statistics, Probability and Uncertainty