No Arabic abstract
In this paper, we investigate an interesting and important stopping problem mixed with stochastic controls and a textit{nonsmooth} utility over a finite time horizon. The paper aims to develop new methodologies, which are significantly different from those of mixed dynamic optimal control and stopping problems in the existing literature, to figure out a managers decision. We formulate our model to a free boundary problem of a fully textit{nonlinear} equation. By means of a dual transformation, however, we can convert the above problem to a new free boundary problem of a textit{linear} equation. Finally, using the corresponding inverse dual transformation, we apply the theoretical results established for the new free boundary problem to obtain the properties of the optimal strategy and the optimal stopping time to achieve a certain level for the original problem over a finite time investment horizon.
In this paper, we investigate dynamic optimization problems featuring both stochastic control and optimal stopping in a finite time horizon. The paper aims to develop new methodologies, which are significantly different from those of mixed dynamic optimal control and stopping problems in the existing literature, to study a managers decision. We formulate our model to a free boundary problem of a fully nonlinear equation. Furthermore, by means of a dual transformation for the above problem, we convert the above problem to a new free boundary problem of a linear equation. Finally, we apply the theoretical results to challenging, yet practically relevant and important, risk-sensitive problems in wealth management to obtain the properties of the optimal strategy and the right time to achieve a certain level over a finite time investment horizon.
In this paper, we consider the optimal stopping problem on semi-Markov processes (SMPs) with finite horizon, and aim to establish the existence and computation of optimal stopping times. To achieve the goal, we first develop the main results of finite horizon semi-Markov decision processes (SMDPs) to the case with additional terminal costs, introduce an explicit construction of SMDPs, and prove the equivalence between the optimal stopping problems on SMPs and SMDPs. Then, using the equivalence and the results on SMDPs developed here, we not only show the existence of optimal stopping time of SMPs, but also provide an algorithm for computing optimal stopping time on SMPs. Moreover, we show that the optimal and -optimal stopping time can be characterized by the hitting time of some special sets, respectively.
We study the optimal investment stopping problem in both continuous and discrete case, where the investor needs to choose the optimal trading strategy and optimal stopping time concurrently to maximize the expected utility of terminal wealth. Based on the work [9] with an additional stochastic payoff function, we characterize the value function for the continuous problem via the theory of quadratic reflected backward stochastic differential equation (BSDE for short) with unbounded terminal condition. In regard to discrete problem, we get the discretization form composed of piecewise quadratic BSDEs recursively under Markovian framework and the assumption of bounded obstacle, and provide some useful prior estimates about the solutions with the help of auxiliary forward-backward SDE system and Malliavin calculus. Finally, we obtain the uniform convergence and relevant rate from discretely to continuously quadratic reflected BSDE, which arise from corresponding optimal investment stopping problem through above characterization.
Intelligent mobile sensors, such as uninhabited aerial or underwater vehicles, are becoming prevalent in environmental sensing and monitoring applications. These active sensing platforms operate in unsteady fluid flows, including windy urban environments, hurricanes, and ocean currents. Often constrained in their actuation capabilities, the dynamics of these mobile sensors depend strongly on the background flow, making their deployment and control particularly challenging. Therefore, efficient trajectory planning with partial knowledge about the background flow is essential for teams of mobile sensors to adaptively sense and monitor their environments. In this work, we investigate the use of finite-horizon model predictive control (MPC) for the energy-efficient trajectory planning of an active mobile sensor in an unsteady fluid flow field. We uncover connections between the finite-time optimal trajectories and finite-time Lyapunov exponents (FTLE) of the background flow, confirming that energy-efficient trajectories exploit invariant coherent structures in the flow. We demonstrate our findings on the unsteady double gyre vector field, which is a canonical model for chaotic mixing in the ocean. We present an exhaustive search through critical MPC parameters including the prediction horizon, maximum sensor actuation, and relative penalty on the accumulated state error and actuation effort. We find that even relatively short prediction horizons can often yield nearly energy-optimal trajectories. These results are promising for the adaptive planning of energy-efficient trajectories for swarms of mobile sensors in distributed sensing and monitoring.
In this paper we study the optimization problem of an economic agent who chooses a job and the time of retirement as well as consumption and portfolio of assets. The agent is constrained in the ability to borrow against future income. We transform the problem into a dual two-person zero-sum game, which involves a controller, who is a minimizer and chooses a non-increasing process, and a stopper, who is a maximizer and chooses a stopping time. We derive the Hamilton-Jacobi- Bellman quasi-variational inequality(HJBQV) of a max-min type arising from the game. We provide a solution to the HJBQV and verification that it is the value of the game. We establish a duality result which allows to derive the optimal strategies and value function of the primal problem from those of the dual problem.