No Arabic abstract
Increasing penetration of wind and renewable generation poses significant challenges to the power system operations and reliability. This paper considers the real-time optimal transmission switching (OTS) problem for determining the generation dispatch and network topology that can account for uncertain energy resources. To efficiently solve the resultant two-stage stochastic program, we propose a tractable linear decision rule (LDR) based approximation solution that can eliminate the uncertainty variables and lead to fixed number of constraints. The proposed LDR approach can guarantee feasibility, and significantly reduces the computational complexity of existing approaches that grows with the number of randomly generated samples of uncertainty. Numerical studies on IEEE test cases demonstrate the high approximation accuracy of the proposed LDR solution and its computational efficiency for real-time OTS implementations.
In this paper we study a class of risk-sensitive Markovian control problems in discrete time subject to model uncertainty. We consider a risk-sensitive discounted cost criterion with finite time horizon. The used methodology is the one of adaptive robust control combined with machine learning.
We study the design of autonomous agents that are capable of deceiving outside observers about their intentions while carrying out tasks in stochastic, complex environments. By modeling the agents behavior as a Markov decision process, we consider a setting where the agent aims to reach one of multiple potential goals while deceiving outside observers about its true goal. We propose a novel approach to model observer predictions based on the principle of maximum entropy and to efficiently generate deceptive strategies via linear programming. The proposed approach enables the agent to exhibit a variety of tunable deceptive behaviors while ensuring the satisfaction of probabilistic constraints on the behavior. We evaluate the performance of the proposed approach via comparative user studies and present a case study on the streets of Manhattan, New York, using real travel time distributions.
Least worst regret (and sometimes minimax) analysis are often used for decision making whenever it is difficult, or inappropriate, to attach probabilities to possible future scenarios. We show that, for each of these two approaches and subject only to the convexity of the cost functions involved, it is always the case that there exist two extreme scenarios whose costs determine the outcome of the analysis in the sense we make clear. The results of either analysis are therefore particularly sensitive to the cost functions associated with these two scenarios, while being largely unaffected by those associated with the remainder. Great care is therefore required in applications to identify these scenarios and to consider their reasonableness. We also consider the relationship between the outcome of a least worst regret and a Bayesian analysis, particularly in the case where the regret functions associated with the scenarios largely differ from each other by shifts in their arguments, as is the case in many applications. We study in detail the problem of determining an appropriate level of electricity capacity procurement in Great Britain, where decisions must be made several years in advance, in spite of considerable uncertainty as to which of a number of future scenarios may occur, and where least worst regret analysis is currently used as the basis of decision making.
We study dynamic allocation problems for discrete time multi-armed bandits under uncertainty, based on the the theory of nonlinear expectations. We show that, under strong independence of the bandits and with some relaxation in the definition of optimality, a Gittins allocation index gives optimal choices. This involves studying the interaction of our uncertainty with controls which determine the filtration. We also run a simple numerical example which illustrates the interaction between the willingness to explore and uncertainty aversion of the agent when making decisions.
We propose a new approach for solving a class of discrete decision making problems under uncertainty with positive cost. This issue concerns multiple and diverse fields such as engineering, economics, artificial intelligence, cognitive science and many others. Basically, an agent has to choose a single or series of actions from a set of options, without knowing for sure their consequences. Schematically, two main approaches have been followed: either the agent learns which option is the correct one to choose in a given situation by trial and error, or the agent already has some knowledge on the possible consequences of his decisions; this knowledge being generally expressed as a conditional probability distribution. In the latter case, several optimal or suboptimal methods have been proposed to exploit this uncertain knowledge in various contexts. In this work, we propose following a different approach, based on the geometric intuition of distance. More precisely, we define a goal independent quasimetric structure on the state space, taking into account both cost function and transition probability. We then compare precision and computation time with classical approaches.