skip to main content


Title: Equilibrium strategies for time-inconsistent stochastic switching systems
An optimal control problem is considered for a stochastic differential equation containing a state-dependent regime switching, with a recursive cost functional. Due to the non-exponential discounting in the cost functional, the problem is time-inconsistent in general. Therefore, instead of finding a global optimal control (which is not possible), we look for a time-consistent (approximately) locally optimal equilibrium strategy. Such a strategy can be represented through the solution to a system of partial differential equations, called an equilibrium Hamilton–Jacob–Bellman (HJB) equation which is constructed via a sequence of multi-person differential games. A verification theorem is proved and, under proper conditions, the well-posedness of the equilibrium HJB equation is established as well.  more » « less
Award ID(s):
1812921
NSF-PAR ID:
10341967
Author(s) / Creator(s):
;
Date Published:
Journal Name:
ESAIM: Control, Optimisation and Calculus of Variations
Volume:
25
ISSN:
1292-8119
Page Range / eLocation ID:
64
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Buttazzo, G. ; Casas, E. ; de Teresa, L. ; Glowinski, R. ; Leugering, G. ; Trélat, E. ; Zhang, X. (Ed.)
    An optimal control problem is considered for a stochastic differential equation with the cost functional determined by a backward stochastic Volterra integral equation (BSVIE, for short). This kind of cost functional can cover the general discounting (including exponential and non-exponential) situations with a recursive feature. It is known that such a problem is time-inconsistent in general. Therefore, instead of finding a global optimal control, we look for a time-consistent locally near optimal equilibrium strategy. With the idea of multi-person differential games, a family of approximate equilibrium strategies is constructed associated with partitions of the time intervals. By sending the mesh size of the time interval partition to zero, an equilibrium Hamilton–Jacobi–Bellman (HJB, for short) equation is derived, through which the equilibrium value function and an equilibrium strategy are obtained. Under certain conditions, a verification theorem is proved and the well-posedness of the equilibrium HJB is established. As a sort of Feynman–Kac formula for the equilibrium HJB equation, a new class of BSVIEs (containing the diagonal value Z ( r , r ) of Z (⋅ , ⋅)) is naturally introduced and the well-posedness of such kind of equations is briefly presented. 
    more » « less
  2. An optimal control problem is considered for linear stochastic differential equations with quadratic cost functional. The coefficients of the state equation and the weights in the cost functional are bounded operators on the spaces of square integrable random variables. The main motivation of our study is linear quadratic (LQ, for short) optimal control problems for mean-field stochastic differential equations. Open-loop solvability of the problem is characterized as the solvability of a system of linear coupled forward-backward stochastic differential equations (FBSDE, for short) with operator coefficients, together with a convexity condition for the cost functional. Under proper conditions, the well-posedness of such an FBSDE, which leads to the existence of an open-loop optimal control, is established. Finally, as applications of our main results, a general mean-field LQ control problem and a concrete mean-variance portfolio selection problem in the open-loop case are solved. 
    more » « less
  3. null (Ed.)
    This paper studies an optimal stochastic impulse control problem in a finite time horizon with a decision lag, by which we mean that after an impulse is made, a fixed number units of time has to be elapsed before the next impulse is allowed to be made. The continuity of the value function is proved. A suitable version of dynamic programming principle is established, which takes into account the dependence of state process on the elapsed time. The corresponding Hamilton-Jacobi-Bellman (HJB) equation is derived, which exhibits some special feature of the problem. The value function of this optimal impulse control problem is characterized as the unique viscosity solution to the corresponding HJB equation. An optimal impulse control is constructed provided the value function is given. Moreover, a limiting case with the waiting time approaching 0 is discussed. 
    more » « less
  4. Abstract

    We develop a continuous‐time control approach to optimal trading in a Proof‐of‐Stake (PoS) blockchain, formulated as a consumption‐investment problem that aims to strike the optimal balance between a participant's (or agent's) utility from holding/trading stakes and utility from consumption. We present solutions via dynamic programming and the Hamilton–Jacobi–Bellman (HJB) equations. When the utility functions are linear or convex, we derive close‐form solutions and show that the bang‐bang strategy is optimal (i.e., always buy or sell at full capacity). Furthermore, we bring out the explicit connection between the rate of return in trading/holding stakes and the participant's risk‐adjusted valuation of the stakes. In particular, we show when a participant is risk‐neutral or risk‐seeking, corresponding to the risk‐adjusted valuation being a martingale or a sub‐martingale, the optimal strategy must be to either buy all the time, sell all the time, or first buy then sell, and with both buying and selling executed at full capacity. We also propose a risk‐control version of the consumption‐investment problem; and for a special case, the “stake‐parity” problem, we show a mean‐reverting strategy is optimal.

     
    more » « less
  5. Introduction

    Runners competing in races are looking to optimize their performance. In this paper, a runner's performance in a race, such as a marathon, is formulated as an optimal control problem where the controls are: the nutrition intake throughout the race and the propulsion force of the runner. As nutrition is an integral part of successfully running long distance races, it needs to be included in models of running strategies.

    Methods

    We formulate a system of ordinary differential equations to represent the velocity, fat energy, glycogen energy, and nutrition for a runner competing in a long-distance race. The energy compartments represent the energy sources available in the runner's body. We allocate the energy source from which the runner draws, based on how fast the runner is moving. The food consumed during the race is a source term for the nutrition differential equation. With our model, we are investigating strategies to manage the nutrition and propulsion force in order to minimize the running time in a fixed distance race. This requires the solution of a nontrivial singular control problem.

    Results

    As the goal of an optimal control model is to determine the optimal strategy, comparing our results against real data presents a challenge; however, in comparing our results to the world record for the marathon, our results differed by 0.4%, 31 seconds. Per each additional gel consumed, the runner is able to run 0.5 to 0.7 kilometers further in the same amount of time, resulting in a 7.75% increase in taking five 100 calorie gels vs no nutrition.

    Discussion

    Our results confirm the belief that the most effective way to run a race is to run approximately the same pace the entire race without letting one's energies hit zero, by consuming in-race nutrition. While this model does not take all factors into account, we consider it a building block for future models, considering our novel energy representation, and in-race nutrition.

     
    more » « less